Mock Version: 3.5 ENTER ['do_with_status'](['bash', '--login', '-c', '/usr/bin/rpmbuild -bs --target x86_64 --nodeps /builddir/build/SPECS/python-scrapy.spec'], chrootPath='/var/lib/mock/rhel+epel-9-x86_64-1683543947.656578/root'env={'TERM': 'vt100', 'SHELL': '/bin/bash', 'HOME': '/builddir', 'HOSTNAME': 'mock', 'PATH': '/usr/bin:/bin:/usr/sbin:/sbin', 'PROMPT_COMMAND': 'printf "\\033]0;\\007"', 'PS1': ' \\s-\\v\\$ ', 'LANG': 'C.UTF-8'}shell=Falselogger=timeout=0uid=1000gid=135user='mockbuild'nspawn_args=['--capability=cap_ipc_lock', '--rlimit=RLIMIT_NOFILE=10240', '--capability=cap_ipc_lock', '--bind=/tmp/mock-resolv.z462kcdf:/etc/resolv.conf', '--bind=/dev/btrfs-control', '--bind=/dev/mapper/control', '--bind=/dev/loop-control', '--bind=/dev/loop0', '--bind=/dev/loop1', '--bind=/dev/loop2', '--bind=/dev/loop3', '--bind=/dev/loop4', '--bind=/dev/loop5', '--bind=/dev/loop6', '--bind=/dev/loop7', '--bind=/dev/loop8', '--bind=/dev/loop9', '--bind=/dev/loop10', '--bind=/dev/loop11']unshare_net=TrueprintOutput=True) Using nspawn with args ['--capability=cap_ipc_lock', '--rlimit=RLIMIT_NOFILE=10240', '--capability=cap_ipc_lock', '--bind=/tmp/mock-resolv.z462kcdf:/etc/resolv.conf', '--bind=/dev/btrfs-control', '--bind=/dev/mapper/control', '--bind=/dev/loop-control', '--bind=/dev/loop0', '--bind=/dev/loop1', '--bind=/dev/loop2', '--bind=/dev/loop3', '--bind=/dev/loop4', '--bind=/dev/loop5', '--bind=/dev/loop6', '--bind=/dev/loop7', '--bind=/dev/loop8', '--bind=/dev/loop9', '--bind=/dev/loop10', '--bind=/dev/loop11'] Executing command: ['/usr/bin/systemd-nspawn', '-q', '-M', '185d8e2c02ab4533953d35dbef115366', '-D', '/var/lib/mock/rhel+epel-9-x86_64-1683543947.656578/root', '-a', '-u', 'mockbuild', '--capability=cap_ipc_lock', '--rlimit=RLIMIT_NOFILE=10240', '--capability=cap_ipc_lock', '--bind=/tmp/mock-resolv.z462kcdf:/etc/resolv.conf', '--bind=/dev/btrfs-control', '--bind=/dev/mapper/control', '--bind=/dev/loop-control', '--bind=/dev/loop0', '--bind=/dev/loop1', '--bind=/dev/loop2', '--bind=/dev/loop3', '--bind=/dev/loop4', '--bind=/dev/loop5', '--bind=/dev/loop6', '--bind=/dev/loop7', '--bind=/dev/loop8', '--bind=/dev/loop9', '--bind=/dev/loop10', '--bind=/dev/loop11', '--console=pipe', '--setenv=TERM=vt100', '--setenv=SHELL=/bin/bash', '--setenv=HOME=/builddir', '--setenv=HOSTNAME=mock', '--setenv=PATH=/usr/bin:/bin:/usr/sbin:/sbin', '--setenv=PROMPT_COMMAND=printf "\\033]0;\\007"', '--setenv=PS1= \\s-\\v\\$ ', '--setenv=LANG=C.UTF-8', '--resolv-conf=off', 'bash', '--login', '-c', '/usr/bin/rpmbuild -bs --target x86_64 --nodeps /builddir/build/SPECS/python-scrapy.spec'] with env {'TERM': 'vt100', 'SHELL': '/bin/bash', 'HOME': '/builddir', 'HOSTNAME': 'mock', 'PATH': '/usr/bin:/bin:/usr/sbin:/sbin', 'PROMPT_COMMAND': 'printf "\\033]0;\\007"', 'PS1': ' \\s-\\v\\$ ', 'LANG': 'C.UTF-8', 'SYSTEMD_NSPAWN_TMPFS_TMP': '0', 'SYSTEMD_SECCOMP': '0'} and shell False Building target platforms: x86_64 Building for target x86_64 setting SOURCE_DATE_EPOCH=1683504000 Wrote: /builddir/build/SRPMS/python-scrapy-2.9.0-1.el9.src.rpm Child return code was: 0 ENTER ['do_with_status'](['bash', '--login', '-c', '/usr/bin/rpmbuild -br --target x86_64 --nodeps /builddir/build/SPECS/python-scrapy.spec'], chrootPath='/var/lib/mock/rhel+epel-9-x86_64-1683543947.656578/root'env={'TERM': 'vt100', 'SHELL': '/bin/bash', 'HOME': '/builddir', 'HOSTNAME': 'mock', 'PATH': '/usr/bin:/bin:/usr/sbin:/sbin', 'PROMPT_COMMAND': 'printf "\\033]0;\\007"', 'PS1': ' \\s-\\v\\$ ', 'LANG': 'C.UTF-8'}shell=Falselogger=timeout=0uid=1000gid=135user='mockbuild'nspawn_args=['--capability=cap_ipc_lock', '--rlimit=RLIMIT_NOFILE=10240', '--capability=cap_ipc_lock', '--bind=/tmp/mock-resolv.z462kcdf:/etc/resolv.conf', '--bind=/dev/btrfs-control', '--bind=/dev/mapper/control', '--bind=/dev/loop-control', '--bind=/dev/loop0', '--bind=/dev/loop1', '--bind=/dev/loop2', '--bind=/dev/loop3', '--bind=/dev/loop4', '--bind=/dev/loop5', '--bind=/dev/loop6', '--bind=/dev/loop7', '--bind=/dev/loop8', '--bind=/dev/loop9', '--bind=/dev/loop10', '--bind=/dev/loop11']unshare_net=TrueraiseExc=FalseprintOutput=True) Using nspawn with args ['--capability=cap_ipc_lock', '--rlimit=RLIMIT_NOFILE=10240', '--capability=cap_ipc_lock', '--bind=/tmp/mock-resolv.z462kcdf:/etc/resolv.conf', '--bind=/dev/btrfs-control', '--bind=/dev/mapper/control', '--bind=/dev/loop-control', '--bind=/dev/loop0', '--bind=/dev/loop1', '--bind=/dev/loop2', '--bind=/dev/loop3', '--bind=/dev/loop4', '--bind=/dev/loop5', '--bind=/dev/loop6', '--bind=/dev/loop7', '--bind=/dev/loop8', '--bind=/dev/loop9', '--bind=/dev/loop10', '--bind=/dev/loop11'] Executing command: ['/usr/bin/systemd-nspawn', '-q', '-M', 'a9f27259d3b2430091372091eb11309c', '-D', '/var/lib/mock/rhel+epel-9-x86_64-1683543947.656578/root', '-a', '-u', 'mockbuild', '--capability=cap_ipc_lock', '--rlimit=RLIMIT_NOFILE=10240', '--capability=cap_ipc_lock', '--bind=/tmp/mock-resolv.z462kcdf:/etc/resolv.conf', '--bind=/dev/btrfs-control', '--bind=/dev/mapper/control', '--bind=/dev/loop-control', '--bind=/dev/loop0', '--bind=/dev/loop1', '--bind=/dev/loop2', '--bind=/dev/loop3', '--bind=/dev/loop4', '--bind=/dev/loop5', '--bind=/dev/loop6', '--bind=/dev/loop7', '--bind=/dev/loop8', '--bind=/dev/loop9', '--bind=/dev/loop10', '--bind=/dev/loop11', '--console=pipe', '--setenv=TERM=vt100', '--setenv=SHELL=/bin/bash', '--setenv=HOME=/builddir', '--setenv=HOSTNAME=mock', '--setenv=PATH=/usr/bin:/bin:/usr/sbin:/sbin', '--setenv=PROMPT_COMMAND=printf "\\033]0;\\007"', '--setenv=PS1= \\s-\\v\\$ ', '--setenv=LANG=C.UTF-8', '--resolv-conf=off', 'bash', '--login', '-c', '/usr/bin/rpmbuild -br --target x86_64 --nodeps /builddir/build/SPECS/python-scrapy.spec'] with env {'TERM': 'vt100', 'SHELL': '/bin/bash', 'HOME': '/builddir', 'HOSTNAME': 'mock', 'PATH': '/usr/bin:/bin:/usr/sbin:/sbin', 'PROMPT_COMMAND': 'printf "\\033]0;\\007"', 'PS1': ' \\s-\\v\\$ ', 'LANG': 'C.UTF-8', 'SYSTEMD_NSPAWN_TMPFS_TMP': '0', 'SYSTEMD_SECCOMP': '0'} and shell False Building target platforms: x86_64 Building for target x86_64 setting SOURCE_DATE_EPOCH=1683504000 Executing(%prep): /bin/sh -e /var/tmp/rpm-tmp.PODTbX + umask 022 + cd /builddir/build/BUILD + cd /builddir/build/BUILD + rm -rf Scrapy-2.9.0 + /usr/bin/gzip -dc /builddir/build/SOURCES/Scrapy-2.9.0.tar.gz + /usr/bin/tar -xof - + STATUS=0 + '[' 0 -ne 0 ']' + cd Scrapy-2.9.0 + /usr/bin/chmod -Rf a+rX,u+w,g-w,o-w . + RPM_EC=0 ++ jobs -p + exit 0 Executing(%generate_buildrequires): /bin/sh -e /var/tmp/rpm-tmp.ltOpmB + umask 022 + cd /builddir/build/BUILD + cd Scrapy-2.9.0 + echo pyproject-rpm-macros + echo python3-devel + echo 'python3dist(pip) >= 19' + echo 'python3dist(packaging)' + '[' -f pyproject.toml ']' + '[' -f setup.py ']' + echo 'python3dist(setuptools) >= 40.8' + echo 'python3dist(wheel)' + rm -rfv '*.dist-info/' + '[' -f /usr/bin/python3 ']' + mkdir -p /builddir/build/BUILD/Scrapy-2.9.0/.pyproject-builddir + CFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -march=x86-64-v2 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection' + LDFLAGS='-Wl,-z,relro -Wl,--as-needed -Wl,-z,now -specs=/usr/lib/rpm/redhat/redhat-hardened-ld -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 ' + TMPDIR=/builddir/build/BUILD/Scrapy-2.9.0/.pyproject-builddir + RPM_TOXENV=py39 + HOSTNAME=rpmbuild + /usr/bin/python3 -Bs /usr/lib/rpm/redhat/pyproject_buildrequires.py --generate-extras --python3_pkgversion 3 --wheeldir /builddir/build/BUILD/Scrapy-2.9.0/pyproject-wheeldir Handling setuptools >= 40.8 from default build backend Requirement satisfied: setuptools >= 40.8 (installed: setuptools 67.7.2) Handling wheel from default build backend Requirement not satisfied: wheel Exiting dependency generation pass: build backend + RPM_EC=0 ++ jobs -p + exit 0 Wrote: /builddir/build/SRPMS/python-scrapy-2.9.0-1.el9.buildreqs.nosrc.rpm Child return code was: 11 Dynamic buildrequires detected Going to install missing buildrequires. See root.log for details. ENTER ['do_with_status'](['bash', '--login', '-c', '/usr/bin/rpmbuild -br --target x86_64 --nodeps /builddir/build/SPECS/python-scrapy.spec'], chrootPath='/var/lib/mock/rhel+epel-9-x86_64-1683543947.656578/root'env={'TERM': 'vt100', 'SHELL': '/bin/bash', 'HOME': '/builddir', 'HOSTNAME': 'mock', 'PATH': '/usr/bin:/bin:/usr/sbin:/sbin', 'PROMPT_COMMAND': 'printf "\\033]0;\\007"', 'PS1': ' \\s-\\v\\$ ', 'LANG': 'C.UTF-8'}shell=Falselogger=timeout=0uid=1000gid=135user='mockbuild'nspawn_args=['--capability=cap_ipc_lock', '--rlimit=RLIMIT_NOFILE=10240', '--capability=cap_ipc_lock', '--bind=/tmp/mock-resolv.z462kcdf:/etc/resolv.conf', '--bind=/dev/btrfs-control', '--bind=/dev/mapper/control', '--bind=/dev/loop-control', '--bind=/dev/loop0', '--bind=/dev/loop1', '--bind=/dev/loop2', '--bind=/dev/loop3', '--bind=/dev/loop4', '--bind=/dev/loop5', '--bind=/dev/loop6', '--bind=/dev/loop7', '--bind=/dev/loop8', '--bind=/dev/loop9', '--bind=/dev/loop10', '--bind=/dev/loop11']unshare_net=TrueraiseExc=FalseprintOutput=True) Using nspawn with args ['--capability=cap_ipc_lock', '--rlimit=RLIMIT_NOFILE=10240', '--capability=cap_ipc_lock', '--bind=/tmp/mock-resolv.z462kcdf:/etc/resolv.conf', '--bind=/dev/btrfs-control', '--bind=/dev/mapper/control', '--bind=/dev/loop-control', '--bind=/dev/loop0', '--bind=/dev/loop1', '--bind=/dev/loop2', '--bind=/dev/loop3', '--bind=/dev/loop4', '--bind=/dev/loop5', '--bind=/dev/loop6', '--bind=/dev/loop7', '--bind=/dev/loop8', '--bind=/dev/loop9', '--bind=/dev/loop10', '--bind=/dev/loop11'] Executing command: ['/usr/bin/systemd-nspawn', '-q', '-M', 'e6379a8464da43eba1562a91ef5792fa', '-D', '/var/lib/mock/rhel+epel-9-x86_64-1683543947.656578/root', '-a', '-u', 'mockbuild', '--capability=cap_ipc_lock', '--rlimit=RLIMIT_NOFILE=10240', '--capability=cap_ipc_lock', '--bind=/tmp/mock-resolv.z462kcdf:/etc/resolv.conf', '--bind=/dev/btrfs-control', '--bind=/dev/mapper/control', '--bind=/dev/loop-control', '--bind=/dev/loop0', '--bind=/dev/loop1', '--bind=/dev/loop2', '--bind=/dev/loop3', '--bind=/dev/loop4', '--bind=/dev/loop5', '--bind=/dev/loop6', '--bind=/dev/loop7', '--bind=/dev/loop8', '--bind=/dev/loop9', '--bind=/dev/loop10', '--bind=/dev/loop11', '--console=pipe', '--setenv=TERM=vt100', '--setenv=SHELL=/bin/bash', '--setenv=HOME=/builddir', '--setenv=HOSTNAME=mock', '--setenv=PATH=/usr/bin:/bin:/usr/sbin:/sbin', '--setenv=PROMPT_COMMAND=printf "\\033]0;\\007"', '--setenv=PS1= \\s-\\v\\$ ', '--setenv=LANG=C.UTF-8', '--resolv-conf=off', 'bash', '--login', '-c', '/usr/bin/rpmbuild -br --target x86_64 --nodeps /builddir/build/SPECS/python-scrapy.spec'] with env {'TERM': 'vt100', 'SHELL': '/bin/bash', 'HOME': '/builddir', 'HOSTNAME': 'mock', 'PATH': '/usr/bin:/bin:/usr/sbin:/sbin', 'PROMPT_COMMAND': 'printf "\\033]0;\\007"', 'PS1': ' \\s-\\v\\$ ', 'LANG': 'C.UTF-8', 'SYSTEMD_NSPAWN_TMPFS_TMP': '0', 'SYSTEMD_SECCOMP': '0'} and shell False Building target platforms: x86_64 Building for target x86_64 setting SOURCE_DATE_EPOCH=1683504000 Executing(%prep): /bin/sh -e /var/tmp/rpm-tmp.3EAazJ + umask 022 + cd /builddir/build/BUILD + cd /builddir/build/BUILD + rm -rf Scrapy-2.9.0 + /usr/bin/gzip -dc /builddir/build/SOURCES/Scrapy-2.9.0.tar.gz + /usr/bin/tar -xof - + STATUS=0 + '[' 0 -ne 0 ']' + cd Scrapy-2.9.0 + /usr/bin/chmod -Rf a+rX,u+w,g-w,o-w . + RPM_EC=0 ++ jobs -p + exit 0 Executing(%generate_buildrequires): /bin/sh -e /var/tmp/rpm-tmp.ZCbudh + umask 022 + cd /builddir/build/BUILD + cd Scrapy-2.9.0 + echo pyproject-rpm-macros + echo python3-devel + echo 'python3dist(pip) >= 19' + echo 'python3dist(packaging)' + '[' -f pyproject.toml ']' + '[' -f setup.py ']' + echo 'python3dist(setuptools) >= 40.8' + echo 'python3dist(wheel)' + rm -rfv '*.dist-info/' + '[' -f /usr/bin/python3 ']' + mkdir -p /builddir/build/BUILD/Scrapy-2.9.0/.pyproject-builddir + CFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -march=x86-64-v2 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection' + LDFLAGS='-Wl,-z,relro -Wl,--as-needed -Wl,-z,now -specs=/usr/lib/rpm/redhat/redhat-hardened-ld -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 ' + TMPDIR=/builddir/build/BUILD/Scrapy-2.9.0/.pyproject-builddir + RPM_TOXENV=py39 + HOSTNAME=rpmbuild + /usr/bin/python3 -Bs /usr/lib/rpm/redhat/pyproject_buildrequires.py --generate-extras --python3_pkgversion 3 --wheeldir /builddir/build/BUILD/Scrapy-2.9.0/pyproject-wheeldir Handling setuptools >= 40.8 from default build backend Requirement satisfied: setuptools >= 40.8 (installed: setuptools 67.7.2) Handling wheel from default build backend Requirement satisfied: wheel (installed: wheel 0.36.2) warning: no files found matching 'INSTALL' warning: no files found matching 'requirements-*.txt' warning: no files found matching 'license.txt' under directory 'scrapy' no previously-included directories found matching 'docs/build' warning: no files found matching '*' under directory 'bin' warning: no previously-included files matching '__pycache__' found anywhere in distribution warning: no previously-included files matching '*.py[cod]' found anywhere in distribution HOOK STDOUT: running egg_info HOOK STDOUT: writing Scrapy.egg-info/PKG-INFO HOOK STDOUT: writing dependency_links to Scrapy.egg-info/dependency_links.txt HOOK STDOUT: writing entry points to Scrapy.egg-info/entry_points.txt HOOK STDOUT: writing requirements to Scrapy.egg-info/requires.txt HOOK STDOUT: writing top-level names to Scrapy.egg-info/top_level.txt HOOK STDOUT: reading manifest file 'Scrapy.egg-info/SOURCES.txt' HOOK STDOUT: reading manifest template 'MANIFEST.in' HOOK STDOUT: adding license file 'LICENSE' HOOK STDOUT: adding license file 'AUTHORS' HOOK STDOUT: writing manifest file 'Scrapy.egg-info/SOURCES.txt' Handling wheel from get_requires_for_build_wheel Requirement satisfied: wheel (installed: wheel 0.36.2) warning: no files found matching 'INSTALL' warning: no files found matching 'requirements-*.txt' warning: no files found matching 'license.txt' under directory 'scrapy' no previously-included directories found matching 'docs/build' warning: no files found matching '*' under directory 'bin' warning: no previously-included files matching '__pycache__' found anywhere in distribution warning: no previously-included files matching '*.py[cod]' found anywhere in distribution Handling Twisted (>=18.9.0) from hook generated metadata: Requires-Dist Requirement not satisfied: Twisted (>=18.9.0) Handling cryptography (>=3.4.6) from hook generated metadata: Requires-Dist Requirement not satisfied: cryptography (>=3.4.6) Handling cssselect (>=0.9.1) from hook generated metadata: Requires-Dist Requirement not satisfied: cssselect (>=0.9.1) Handling itemloaders (>=1.0.1) from hook generated metadata: Requires-Dist Requirement not satisfied: itemloaders (>=1.0.1) Handling parsel (>=1.5.0) from hook generated metadata: Requires-Dist Requirement not satisfied: parsel (>=1.5.0) Handling pyOpenSSL (>=21.0.0) from hook generated metadata: Requires-Dist Requirement not satisfied: pyOpenSSL (>=21.0.0) Handling queuelib (>=1.4.2) from hook generated metadata: Requires-Dist Requirement not satisfied: queuelib (>=1.4.2) Handling service-identity (>=18.1.0) from hook generated metadata: Requires-Dist Requirement not satisfied: service-identity (>=18.1.0) Handling w3lib (>=1.17.0) from hook generated metadata: Requires-Dist Requirement not satisfied: w3lib (>=1.17.0) Handling zope.interface (>=5.1.0) from hook generated metadata: Requires-Dist Requirement not satisfied: zope.interface (>=5.1.0) Handling protego (>=0.1.15) from hook generated metadata: Requires-Dist Requirement not satisfied: protego (>=0.1.15) Handling itemadapter (>=0.1.0) from hook generated metadata: Requires-Dist Requirement not satisfied: itemadapter (>=0.1.0) Handling setuptools from hook generated metadata: Requires-Dist Requirement satisfied: setuptools (installed: setuptools 67.7.2) Handling packaging from hook generated metadata: Requires-Dist Requirement satisfied: packaging (installed: packaging 23.1) Handling tldextract from hook generated metadata: Requires-Dist Requirement not satisfied: tldextract Handling lxml (>=4.3.0) from hook generated metadata: Requires-Dist Requirement not satisfied: lxml (>=4.3.0) Handling PyDispatcher (>=2.0.5) ; platform_python_implementation == "CPython" from hook generated metadata: Requires-Dist Requirement not satisfied: PyDispatcher (>=2.0.5) ; platform_python_implementation == "CPython" Handling PyPyDispatcher (>=2.1.0) ; platform_python_implementation == "PyPy" from hook generated metadata: Requires-Dist Ignoring alien requirement: PyPyDispatcher (>=2.1.0) ; platform_python_implementation == "PyPy" + RPM_EC=0 ++ jobs -p + exit 0 Wrote: /builddir/build/SRPMS/python-scrapy-2.9.0-1.el9.buildreqs.nosrc.rpm Child return code was: 11 Dynamic buildrequires detected Going to install missing buildrequires. See root.log for details. ENTER ['do_with_status'](['bash', '--login', '-c', '/usr/bin/rpmbuild -br --target x86_64 --nodeps /builddir/build/SPECS/python-scrapy.spec'], chrootPath='/var/lib/mock/rhel+epel-9-x86_64-1683543947.656578/root'env={'TERM': 'vt100', 'SHELL': '/bin/bash', 'HOME': '/builddir', 'HOSTNAME': 'mock', 'PATH': '/usr/bin:/bin:/usr/sbin:/sbin', 'PROMPT_COMMAND': 'printf "\\033]0;\\007"', 'PS1': ' \\s-\\v\\$ ', 'LANG': 'C.UTF-8'}shell=Falselogger=timeout=0uid=1000gid=135user='mockbuild'nspawn_args=['--capability=cap_ipc_lock', '--rlimit=RLIMIT_NOFILE=10240', '--capability=cap_ipc_lock', '--bind=/tmp/mock-resolv.z462kcdf:/etc/resolv.conf', '--bind=/dev/btrfs-control', '--bind=/dev/mapper/control', '--bind=/dev/loop-control', '--bind=/dev/loop0', '--bind=/dev/loop1', '--bind=/dev/loop2', '--bind=/dev/loop3', '--bind=/dev/loop4', '--bind=/dev/loop5', '--bind=/dev/loop6', '--bind=/dev/loop7', '--bind=/dev/loop8', '--bind=/dev/loop9', '--bind=/dev/loop10', '--bind=/dev/loop11']unshare_net=TrueraiseExc=FalseprintOutput=True) Using nspawn with args ['--capability=cap_ipc_lock', '--rlimit=RLIMIT_NOFILE=10240', '--capability=cap_ipc_lock', '--bind=/tmp/mock-resolv.z462kcdf:/etc/resolv.conf', '--bind=/dev/btrfs-control', '--bind=/dev/mapper/control', '--bind=/dev/loop-control', '--bind=/dev/loop0', '--bind=/dev/loop1', '--bind=/dev/loop2', '--bind=/dev/loop3', '--bind=/dev/loop4', '--bind=/dev/loop5', '--bind=/dev/loop6', '--bind=/dev/loop7', '--bind=/dev/loop8', '--bind=/dev/loop9', '--bind=/dev/loop10', '--bind=/dev/loop11'] Executing command: ['/usr/bin/systemd-nspawn', '-q', '-M', 'f352195c41f547fb9e1cf450edf12b4b', '-D', '/var/lib/mock/rhel+epel-9-x86_64-1683543947.656578/root', '-a', '-u', 'mockbuild', '--capability=cap_ipc_lock', '--rlimit=RLIMIT_NOFILE=10240', '--capability=cap_ipc_lock', '--bind=/tmp/mock-resolv.z462kcdf:/etc/resolv.conf', '--bind=/dev/btrfs-control', '--bind=/dev/mapper/control', '--bind=/dev/loop-control', '--bind=/dev/loop0', '--bind=/dev/loop1', '--bind=/dev/loop2', '--bind=/dev/loop3', '--bind=/dev/loop4', '--bind=/dev/loop5', '--bind=/dev/loop6', '--bind=/dev/loop7', '--bind=/dev/loop8', '--bind=/dev/loop9', '--bind=/dev/loop10', '--bind=/dev/loop11', '--console=pipe', '--setenv=TERM=vt100', '--setenv=SHELL=/bin/bash', '--setenv=HOME=/builddir', '--setenv=HOSTNAME=mock', '--setenv=PATH=/usr/bin:/bin:/usr/sbin:/sbin', '--setenv=PROMPT_COMMAND=printf "\\033]0;\\007"', '--setenv=PS1= \\s-\\v\\$ ', '--setenv=LANG=C.UTF-8', '--resolv-conf=off', 'bash', '--login', '-c', '/usr/bin/rpmbuild -br --target x86_64 --nodeps /builddir/build/SPECS/python-scrapy.spec'] with env {'TERM': 'vt100', 'SHELL': '/bin/bash', 'HOME': '/builddir', 'HOSTNAME': 'mock', 'PATH': '/usr/bin:/bin:/usr/sbin:/sbin', 'PROMPT_COMMAND': 'printf "\\033]0;\\007"', 'PS1': ' \\s-\\v\\$ ', 'LANG': 'C.UTF-8', 'SYSTEMD_NSPAWN_TMPFS_TMP': '0', 'SYSTEMD_SECCOMP': '0'} and shell False Building target platforms: x86_64 Building for target x86_64 setting SOURCE_DATE_EPOCH=1683504000 Executing(%prep): /bin/sh -e /var/tmp/rpm-tmp.rgmMQ7 + umask 022 + cd /builddir/build/BUILD + cd /builddir/build/BUILD + rm -rf Scrapy-2.9.0 + /usr/bin/gzip -dc /builddir/build/SOURCES/Scrapy-2.9.0.tar.gz + /usr/bin/tar -xof - + STATUS=0 + '[' 0 -ne 0 ']' + cd Scrapy-2.9.0 + /usr/bin/chmod -Rf a+rX,u+w,g-w,o-w . + RPM_EC=0 ++ jobs -p + exit 0 Executing(%generate_buildrequires): /bin/sh -e /var/tmp/rpm-tmp.UUmhMU + umask 022 + cd /builddir/build/BUILD + cd Scrapy-2.9.0 + echo pyproject-rpm-macros + echo python3-devel + echo 'python3dist(pip) >= 19' + echo 'python3dist(packaging)' + '[' -f pyproject.toml ']' + '[' -f setup.py ']' + echo 'python3dist(setuptools) >= 40.8' + echo 'python3dist(wheel)' + rm -rfv '*.dist-info/' + '[' -f /usr/bin/python3 ']' + mkdir -p /builddir/build/BUILD/Scrapy-2.9.0/.pyproject-builddir + CFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -march=x86-64-v2 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection' + LDFLAGS='-Wl,-z,relro -Wl,--as-needed -Wl,-z,now -specs=/usr/lib/rpm/redhat/redhat-hardened-ld -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 ' + TMPDIR=/builddir/build/BUILD/Scrapy-2.9.0/.pyproject-builddir + RPM_TOXENV=py39 + HOSTNAME=rpmbuild + /usr/bin/python3 -Bs /usr/lib/rpm/redhat/pyproject_buildrequires.py --generate-extras --python3_pkgversion 3 --wheeldir /builddir/build/BUILD/Scrapy-2.9.0/pyproject-wheeldir Handling setuptools >= 40.8 from default build backend Requirement satisfied: setuptools >= 40.8 (installed: setuptools 67.7.2) Handling wheel from default build backend Requirement satisfied: wheel (installed: wheel 0.36.2) warning: no files found matching 'INSTALL' warning: no files found matching 'requirements-*.txt' warning: no files found matching 'license.txt' under directory 'scrapy' no previously-included directories found matching 'docs/build' warning: no files found matching '*' under directory 'bin' warning: no previously-included files matching '__pycache__' found anywhere in distribution warning: no previously-included files matching '*.py[cod]' found anywhere in distribution HOOK STDOUT: running egg_info HOOK STDOUT: writing Scrapy.egg-info/PKG-INFO HOOK STDOUT: writing dependency_links to Scrapy.egg-info/dependency_links.txt HOOK STDOUT: writing entry points to Scrapy.egg-info/entry_points.txt HOOK STDOUT: writing requirements to Scrapy.egg-info/requires.txt HOOK STDOUT: writing top-level names to Scrapy.egg-info/top_level.txt HOOK STDOUT: reading manifest file 'Scrapy.egg-info/SOURCES.txt' HOOK STDOUT: reading manifest template 'MANIFEST.in' HOOK STDOUT: adding license file 'LICENSE' HOOK STDOUT: adding license file 'AUTHORS' HOOK STDOUT: writing manifest file 'Scrapy.egg-info/SOURCES.txt' Handling wheel from get_requires_for_build_wheel Requirement satisfied: wheel (installed: wheel 0.36.2) warning: no files found matching 'INSTALL' warning: no files found matching 'requirements-*.txt' warning: no files found matching 'license.txt' under directory 'scrapy' no previously-included directories found matching 'docs/build' warning: no files found matching '*' under directory 'bin' warning: no previously-included files matching '__pycache__' found anywhere in distribution warning: no previously-included files matching '*.py[cod]' found anywhere in distribution Handling Twisted (>=18.9.0) from hook generated metadata: Requires-Dist Requirement satisfied: Twisted (>=18.9.0) (installed: Twisted 22.10.0) Handling cryptography (>=3.4.6) from hook generated metadata: Requires-Dist Requirement satisfied: cryptography (>=3.4.6) (installed: cryptography 36.0.1) Handling cssselect (>=0.9.1) from hook generated metadata: Requires-Dist Requirement satisfied: cssselect (>=0.9.1) (installed: cssselect 1.1.0) Handling itemloaders (>=1.0.1) from hook generated metadata: Requires-Dist Requirement satisfied: itemloaders (>=1.0.1) (installed: itemloaders 1.1.0) Handling parsel (>=1.5.0) from hook generated metadata: Requires-Dist Requirement satisfied: parsel (>=1.5.0) (installed: parsel 1.7.0) Handling pyOpenSSL (>=21.0.0) from hook generated metadata: Requires-Dist Requirement satisfied: pyOpenSSL (>=21.0.0) (installed: pyOpenSSL 22.0.0) Handling queuelib (>=1.4.2) from hook generated metadata: Requires-Dist Requirement satisfied: queuelib (>=1.4.2) (installed: queuelib 1.6.2) Handling service-identity (>=18.1.0) from hook generated metadata: Requires-Dist Requirement satisfied: service-identity (>=18.1.0) (installed: service-identity 21.1.0) Handling w3lib (>=1.17.0) from hook generated metadata: Requires-Dist Requirement satisfied: w3lib (>=1.17.0) (installed: w3lib 2.0.1) Handling zope.interface (>=5.1.0) from hook generated metadata: Requires-Dist Requirement satisfied: zope.interface (>=5.1.0) (installed: zope.interface 5.4.0) Handling protego (>=0.1.15) from hook generated metadata: Requires-Dist Requirement satisfied: protego (>=0.1.15) (installed: protego 0.2.1) Handling itemadapter (>=0.1.0) from hook generated metadata: Requires-Dist Requirement satisfied: itemadapter (>=0.1.0) (installed: itemadapter 0.8.0) Handling setuptools from hook generated metadata: Requires-Dist Requirement satisfied: setuptools (installed: setuptools 67.7.2) Handling packaging from hook generated metadata: Requires-Dist Requirement satisfied: packaging (installed: packaging 23.1) Handling tldextract from hook generated metadata: Requires-Dist Requirement satisfied: tldextract (installed: tldextract 3.4.1) Handling lxml (>=4.3.0) from hook generated metadata: Requires-Dist Requirement satisfied: lxml (>=4.3.0) (installed: lxml 4.6.5) Handling PyDispatcher (>=2.0.5) ; platform_python_implementation == "CPython" from hook generated metadata: Requires-Dist Requirement satisfied: PyDispatcher (>=2.0.5) ; platform_python_implementation == "CPython" (installed: PyDispatcher 2.0.7) Handling PyPyDispatcher (>=2.1.0) ; platform_python_implementation == "PyPy" from hook generated metadata: Requires-Dist Ignoring alien requirement: PyPyDispatcher (>=2.1.0) ; platform_python_implementation == "PyPy" + RPM_EC=0 ++ jobs -p + exit 0 Wrote: /builddir/build/SRPMS/python-scrapy-2.9.0-1.el9.buildreqs.nosrc.rpm Child return code was: 11 Dynamic buildrequires detected Going to install missing buildrequires. See root.log for details. ENTER ['do_with_status'](['bash', '--login', '-c', '/usr/bin/rpmbuild -ba --noprep --target x86_64 --nodeps /builddir/build/SPECS/python-scrapy.spec'], chrootPath='/var/lib/mock/rhel+epel-9-x86_64-1683543947.656578/root'env={'TERM': 'vt100', 'SHELL': '/bin/bash', 'HOME': '/builddir', 'HOSTNAME': 'mock', 'PATH': '/usr/bin:/bin:/usr/sbin:/sbin', 'PROMPT_COMMAND': 'printf "\\033]0;\\007"', 'PS1': ' \\s-\\v\\$ ', 'LANG': 'C.UTF-8'}shell=Falselogger=timeout=0uid=1000gid=135user='mockbuild'nspawn_args=['--capability=cap_ipc_lock', '--rlimit=RLIMIT_NOFILE=10240', '--capability=cap_ipc_lock', '--bind=/tmp/mock-resolv.z462kcdf:/etc/resolv.conf', '--bind=/dev/btrfs-control', '--bind=/dev/mapper/control', '--bind=/dev/loop-control', '--bind=/dev/loop0', '--bind=/dev/loop1', '--bind=/dev/loop2', '--bind=/dev/loop3', '--bind=/dev/loop4', '--bind=/dev/loop5', '--bind=/dev/loop6', '--bind=/dev/loop7', '--bind=/dev/loop8', '--bind=/dev/loop9', '--bind=/dev/loop10', '--bind=/dev/loop11']unshare_net=TrueprintOutput=True) Using nspawn with args ['--capability=cap_ipc_lock', '--rlimit=RLIMIT_NOFILE=10240', '--capability=cap_ipc_lock', '--bind=/tmp/mock-resolv.z462kcdf:/etc/resolv.conf', '--bind=/dev/btrfs-control', '--bind=/dev/mapper/control', '--bind=/dev/loop-control', '--bind=/dev/loop0', '--bind=/dev/loop1', '--bind=/dev/loop2', '--bind=/dev/loop3', '--bind=/dev/loop4', '--bind=/dev/loop5', '--bind=/dev/loop6', '--bind=/dev/loop7', '--bind=/dev/loop8', '--bind=/dev/loop9', '--bind=/dev/loop10', '--bind=/dev/loop11'] Executing command: ['/usr/bin/systemd-nspawn', '-q', '-M', 'f417024fc64e43e898871a60ac18b694', '-D', '/var/lib/mock/rhel+epel-9-x86_64-1683543947.656578/root', '-a', '-u', 'mockbuild', '--capability=cap_ipc_lock', '--rlimit=RLIMIT_NOFILE=10240', '--capability=cap_ipc_lock', '--bind=/tmp/mock-resolv.z462kcdf:/etc/resolv.conf', '--bind=/dev/btrfs-control', '--bind=/dev/mapper/control', '--bind=/dev/loop-control', '--bind=/dev/loop0', '--bind=/dev/loop1', '--bind=/dev/loop2', '--bind=/dev/loop3', '--bind=/dev/loop4', '--bind=/dev/loop5', '--bind=/dev/loop6', '--bind=/dev/loop7', '--bind=/dev/loop8', '--bind=/dev/loop9', '--bind=/dev/loop10', '--bind=/dev/loop11', '--console=pipe', '--setenv=TERM=vt100', '--setenv=SHELL=/bin/bash', '--setenv=HOME=/builddir', '--setenv=HOSTNAME=mock', '--setenv=PATH=/usr/bin:/bin:/usr/sbin:/sbin', '--setenv=PROMPT_COMMAND=printf "\\033]0;\\007"', '--setenv=PS1= \\s-\\v\\$ ', '--setenv=LANG=C.UTF-8', '--resolv-conf=off', 'bash', '--login', '-c', '/usr/bin/rpmbuild -ba --noprep --target x86_64 --nodeps /builddir/build/SPECS/python-scrapy.spec'] with env {'TERM': 'vt100', 'SHELL': '/bin/bash', 'HOME': '/builddir', 'HOSTNAME': 'mock', 'PATH': '/usr/bin:/bin:/usr/sbin:/sbin', 'PROMPT_COMMAND': 'printf "\\033]0;\\007"', 'PS1': ' \\s-\\v\\$ ', 'LANG': 'C.UTF-8', 'SYSTEMD_NSPAWN_TMPFS_TMP': '0', 'SYSTEMD_SECCOMP': '0'} and shell False Building target platforms: x86_64 Building for target x86_64 setting SOURCE_DATE_EPOCH=1683504000 Executing(%generate_buildrequires): /bin/sh -e /var/tmp/rpm-tmp.vsPKI8 + umask 022 + cd /builddir/build/BUILD + cd Scrapy-2.9.0 + echo pyproject-rpm-macros + echo python3-devel + echo 'python3dist(pip) >= 19' + echo 'python3dist(packaging)' + '[' -f pyproject.toml ']' + '[' -f setup.py ']' + echo 'python3dist(setuptools) >= 40.8' + echo 'python3dist(wheel)' + rm -rfv Scrapy-2.9.0.dist-info/ removed 'Scrapy-2.9.0.dist-info/AUTHORS' removed 'Scrapy-2.9.0.dist-info/LICENSE' removed 'Scrapy-2.9.0.dist-info/METADATA' removed 'Scrapy-2.9.0.dist-info/entry_points.txt' removed 'Scrapy-2.9.0.dist-info/top_level.txt' removed directory 'Scrapy-2.9.0.dist-info/' + '[' -f /usr/bin/python3 ']' + mkdir -p /builddir/build/BUILD/Scrapy-2.9.0/.pyproject-builddir + CFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -march=x86-64-v2 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection' + LDFLAGS='-Wl,-z,relro -Wl,--as-needed -Wl,-z,now -specs=/usr/lib/rpm/redhat/redhat-hardened-ld -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 ' + TMPDIR=/builddir/build/BUILD/Scrapy-2.9.0/.pyproject-builddir + RPM_TOXENV=py39 + HOSTNAME=rpmbuild + /usr/bin/python3 -Bs /usr/lib/rpm/redhat/pyproject_buildrequires.py --generate-extras --python3_pkgversion 3 --wheeldir /builddir/build/BUILD/Scrapy-2.9.0/pyproject-wheeldir Handling setuptools >= 40.8 from default build backend Requirement satisfied: setuptools >= 40.8 (installed: setuptools 67.7.2) Handling wheel from default build backend Requirement satisfied: wheel (installed: wheel 0.36.2) warning: no files found matching 'INSTALL' warning: no files found matching 'requirements-*.txt' warning: no files found matching 'license.txt' under directory 'scrapy' no previously-included directories found matching 'docs/build' warning: no files found matching '*' under directory 'bin' warning: no previously-included files matching '__pycache__' found anywhere in distribution warning: no previously-included files matching '*.py[cod]' found anywhere in distribution HOOK STDOUT: running egg_info HOOK STDOUT: writing Scrapy.egg-info/PKG-INFO HOOK STDOUT: writing dependency_links to Scrapy.egg-info/dependency_links.txt HOOK STDOUT: writing entry points to Scrapy.egg-info/entry_points.txt HOOK STDOUT: writing requirements to Scrapy.egg-info/requires.txt HOOK STDOUT: writing top-level names to Scrapy.egg-info/top_level.txt HOOK STDOUT: reading manifest file 'Scrapy.egg-info/SOURCES.txt' HOOK STDOUT: reading manifest template 'MANIFEST.in' HOOK STDOUT: adding license file 'LICENSE' HOOK STDOUT: adding license file 'AUTHORS' HOOK STDOUT: writing manifest file 'Scrapy.egg-info/SOURCES.txt' Handling wheel from get_requires_for_build_wheel Requirement satisfied: wheel (installed: wheel 0.36.2) warning: no files found matching 'INSTALL' warning: no files found matching 'requirements-*.txt' warning: no files found matching 'license.txt' under directory 'scrapy' no previously-included directories found matching 'docs/build' warning: no files found matching '*' under directory 'bin' warning: no previously-included files matching '__pycache__' found anywhere in distribution warning: no previously-included files matching '*.py[cod]' found anywhere in distribution Handling Twisted (>=18.9.0) from hook generated metadata: Requires-Dist Requirement satisfied: Twisted (>=18.9.0) (installed: Twisted 22.10.0) Handling cryptography (>=3.4.6) from hook generated metadata: Requires-Dist Requirement satisfied: cryptography (>=3.4.6) (installed: cryptography 36.0.1) Handling cssselect (>=0.9.1) from hook generated metadata: Requires-Dist Requirement satisfied: cssselect (>=0.9.1) (installed: cssselect 1.1.0) Handling itemloaders (>=1.0.1) from hook generated metadata: Requires-Dist Requirement satisfied: itemloaders (>=1.0.1) (installed: itemloaders 1.1.0) Handling parsel (>=1.5.0) from hook generated metadata: Requires-Dist Requirement satisfied: parsel (>=1.5.0) (installed: parsel 1.7.0) Handling pyOpenSSL (>=21.0.0) from hook generated metadata: Requires-Dist Requirement satisfied: pyOpenSSL (>=21.0.0) (installed: pyOpenSSL 22.0.0) Handling queuelib (>=1.4.2) from hook generated metadata: Requires-Dist Requirement satisfied: queuelib (>=1.4.2) (installed: queuelib 1.6.2) Handling service-identity (>=18.1.0) from hook generated metadata: Requires-Dist Requirement satisfied: service-identity (>=18.1.0) (installed: service-identity 21.1.0) Handling w3lib (>=1.17.0) from hook generated metadata: Requires-Dist Requirement satisfied: w3lib (>=1.17.0) (installed: w3lib 2.0.1) Handling zope.interface (>=5.1.0) from hook generated metadata: Requires-Dist Requirement satisfied: zope.interface (>=5.1.0) (installed: zope.interface 5.4.0) Handling protego (>=0.1.15) from hook generated metadata: Requires-Dist Requirement satisfied: protego (>=0.1.15) (installed: protego 0.2.1) Handling itemadapter (>=0.1.0) from hook generated metadata: Requires-Dist Requirement satisfied: itemadapter (>=0.1.0) (installed: itemadapter 0.8.0) Handling setuptools from hook generated metadata: Requires-Dist Requirement satisfied: setuptools (installed: setuptools 67.7.2) Handling packaging from hook generated metadata: Requires-Dist Requirement satisfied: packaging (installed: packaging 23.1) Handling tldextract from hook generated metadata: Requires-Dist Requirement satisfied: tldextract (installed: tldextract 3.4.1) Handling lxml (>=4.3.0) from hook generated metadata: Requires-Dist Requirement satisfied: lxml (>=4.3.0) (installed: lxml 4.6.5) Handling PyDispatcher (>=2.0.5) ; platform_python_implementation == "CPython" from hook generated metadata: Requires-Dist Requirement satisfied: PyDispatcher (>=2.0.5) ; platform_python_implementation == "CPython" (installed: PyDispatcher 2.0.7) Handling PyPyDispatcher (>=2.1.0) ; platform_python_implementation == "PyPy" from hook generated metadata: Requires-Dist Ignoring alien requirement: PyPyDispatcher (>=2.1.0) ; platform_python_implementation == "PyPy" + RPM_EC=0 ++ jobs -p + exit 0 Executing(%build): /bin/sh -e /var/tmp/rpm-tmp.Hysi6n + umask 022 + cd /builddir/build/BUILD + cd Scrapy-2.9.0 + mkdir -p /builddir/build/BUILD/Scrapy-2.9.0/.pyproject-builddir + CFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -march=x86-64-v2 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection' + LDFLAGS='-Wl,-z,relro -Wl,--as-needed -Wl,-z,now -specs=/usr/lib/rpm/redhat/redhat-hardened-ld -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 ' + TMPDIR=/builddir/build/BUILD/Scrapy-2.9.0/.pyproject-builddir + /usr/bin/python3 -Bs /usr/lib/rpm/redhat/pyproject_wheel.py /builddir/build/BUILD/Scrapy-2.9.0/pyproject-wheeldir Processing /builddir/build/BUILD/Scrapy-2.9.0 Preparing metadata (pyproject.toml): started Running command Preparing metadata (pyproject.toml) running dist_info creating /builddir/build/BUILD/Scrapy-2.9.0/.pyproject-builddir/pip-modern-metadata-2ob51k6e/Scrapy.egg-info writing /builddir/build/BUILD/Scrapy-2.9.0/.pyproject-builddir/pip-modern-metadata-2ob51k6e/Scrapy.egg-info/PKG-INFO writing dependency_links to /builddir/build/BUILD/Scrapy-2.9.0/.pyproject-builddir/pip-modern-metadata-2ob51k6e/Scrapy.egg-info/dependency_links.txt writing entry points to /builddir/build/BUILD/Scrapy-2.9.0/.pyproject-builddir/pip-modern-metadata-2ob51k6e/Scrapy.egg-info/entry_points.txt writing requirements to /builddir/build/BUILD/Scrapy-2.9.0/.pyproject-builddir/pip-modern-metadata-2ob51k6e/Scrapy.egg-info/requires.txt writing top-level names to /builddir/build/BUILD/Scrapy-2.9.0/.pyproject-builddir/pip-modern-metadata-2ob51k6e/Scrapy.egg-info/top_level.txt writing manifest file '/builddir/build/BUILD/Scrapy-2.9.0/.pyproject-builddir/pip-modern-metadata-2ob51k6e/Scrapy.egg-info/SOURCES.txt' reading manifest file '/builddir/build/BUILD/Scrapy-2.9.0/.pyproject-builddir/pip-modern-metadata-2ob51k6e/Scrapy.egg-info/SOURCES.txt' reading manifest template 'MANIFEST.in' warning: no files found matching 'INSTALL' warning: no files found matching 'requirements-*.txt' warning: no files found matching 'license.txt' under directory 'scrapy' no previously-included directories found matching 'docs/build' warning: no files found matching '*' under directory 'bin' warning: no previously-included files matching '__pycache__' found anywhere in distribution warning: no previously-included files matching '*.py[cod]' found anywhere in distribution adding license file 'LICENSE' adding license file 'AUTHORS' writing manifest file '/builddir/build/BUILD/Scrapy-2.9.0/.pyproject-builddir/pip-modern-metadata-2ob51k6e/Scrapy.egg-info/SOURCES.txt' creating '/builddir/build/BUILD/Scrapy-2.9.0/.pyproject-builddir/pip-modern-metadata-2ob51k6e/Scrapy-2.9.0.dist-info' adding license file "LICENSE" (matched pattern "LICEN[CS]E*") adding license file "AUTHORS" (matched pattern "AUTHORS*") Preparing metadata (pyproject.toml): finished with status 'done' Building wheels for collected packages: Scrapy Building wheel for Scrapy (pyproject.toml): started Running command Building wheel for Scrapy (pyproject.toml) running bdist_wheel running build running build_py creating build creating build/lib creating build/lib/scrapy copying scrapy/statscollectors.py -> build/lib/scrapy copying scrapy/squeues.py -> build/lib/scrapy copying scrapy/spiderloader.py -> build/lib/scrapy copying scrapy/signals.py -> build/lib/scrapy copying scrapy/signalmanager.py -> build/lib/scrapy copying scrapy/shell.py -> build/lib/scrapy copying scrapy/robotstxt.py -> build/lib/scrapy copying scrapy/responsetypes.py -> build/lib/scrapy copying scrapy/resolver.py -> build/lib/scrapy copying scrapy/pqueues.py -> build/lib/scrapy copying scrapy/middleware.py -> build/lib/scrapy copying scrapy/mail.py -> build/lib/scrapy copying scrapy/logformatter.py -> build/lib/scrapy copying scrapy/link.py -> build/lib/scrapy copying scrapy/item.py -> build/lib/scrapy copying scrapy/interfaces.py -> build/lib/scrapy copying scrapy/extension.py -> build/lib/scrapy copying scrapy/exporters.py -> build/lib/scrapy copying scrapy/exceptions.py -> build/lib/scrapy copying scrapy/dupefilters.py -> build/lib/scrapy copying scrapy/crawler.py -> build/lib/scrapy copying scrapy/cmdline.py -> build/lib/scrapy copying scrapy/__main__.py -> build/lib/scrapy copying scrapy/__init__.py -> build/lib/scrapy creating build/lib/scrapy/utils copying scrapy/utils/versions.py -> build/lib/scrapy/utils copying scrapy/utils/url.py -> build/lib/scrapy/utils copying scrapy/utils/trackref.py -> build/lib/scrapy/utils copying scrapy/utils/testsite.py -> build/lib/scrapy/utils copying scrapy/utils/testproc.py -> build/lib/scrapy/utils copying scrapy/utils/test.py -> build/lib/scrapy/utils copying scrapy/utils/template.py -> build/lib/scrapy/utils copying scrapy/utils/ssl.py -> build/lib/scrapy/utils copying scrapy/utils/spider.py -> build/lib/scrapy/utils copying scrapy/utils/sitemap.py -> build/lib/scrapy/utils copying scrapy/utils/signal.py -> build/lib/scrapy/utils copying scrapy/utils/serialize.py -> build/lib/scrapy/utils copying scrapy/utils/response.py -> build/lib/scrapy/utils copying scrapy/utils/request.py -> build/lib/scrapy/utils copying scrapy/utils/reqser.py -> build/lib/scrapy/utils copying scrapy/utils/reactor.py -> build/lib/scrapy/utils copying scrapy/utils/python.py -> build/lib/scrapy/utils copying scrapy/utils/project.py -> build/lib/scrapy/utils copying scrapy/utils/ossignal.py -> build/lib/scrapy/utils copying scrapy/utils/misc.py -> build/lib/scrapy/utils copying scrapy/utils/log.py -> build/lib/scrapy/utils copying scrapy/utils/job.py -> build/lib/scrapy/utils copying scrapy/utils/iterators.py -> build/lib/scrapy/utils copying scrapy/utils/httpobj.py -> build/lib/scrapy/utils copying scrapy/utils/gz.py -> build/lib/scrapy/utils copying scrapy/utils/ftp.py -> build/lib/scrapy/utils copying scrapy/utils/engine.py -> build/lib/scrapy/utils copying scrapy/utils/display.py -> build/lib/scrapy/utils copying scrapy/utils/deprecate.py -> build/lib/scrapy/utils copying scrapy/utils/defer.py -> build/lib/scrapy/utils copying scrapy/utils/decorators.py -> build/lib/scrapy/utils copying scrapy/utils/datatypes.py -> build/lib/scrapy/utils copying scrapy/utils/curl.py -> build/lib/scrapy/utils copying scrapy/utils/console.py -> build/lib/scrapy/utils copying scrapy/utils/conf.py -> build/lib/scrapy/utils copying scrapy/utils/boto.py -> build/lib/scrapy/utils copying scrapy/utils/benchserver.py -> build/lib/scrapy/utils copying scrapy/utils/asyncgen.py -> build/lib/scrapy/utils copying scrapy/utils/__init__.py -> build/lib/scrapy/utils creating build/lib/scrapy/spiders copying scrapy/spiders/sitemap.py -> build/lib/scrapy/spiders copying scrapy/spiders/init.py -> build/lib/scrapy/spiders copying scrapy/spiders/feed.py -> build/lib/scrapy/spiders copying scrapy/spiders/crawl.py -> build/lib/scrapy/spiders copying scrapy/spiders/__init__.py -> build/lib/scrapy/spiders creating build/lib/scrapy/spidermiddlewares copying scrapy/spidermiddlewares/urllength.py -> build/lib/scrapy/spidermiddlewares copying scrapy/spidermiddlewares/referer.py -> build/lib/scrapy/spidermiddlewares copying scrapy/spidermiddlewares/offsite.py -> build/lib/scrapy/spidermiddlewares copying scrapy/spidermiddlewares/httperror.py -> build/lib/scrapy/spidermiddlewares copying scrapy/spidermiddlewares/depth.py -> build/lib/scrapy/spidermiddlewares copying scrapy/spidermiddlewares/__init__.py -> build/lib/scrapy/spidermiddlewares creating build/lib/scrapy/settings copying scrapy/settings/default_settings.py -> build/lib/scrapy/settings copying scrapy/settings/__init__.py -> build/lib/scrapy/settings creating build/lib/scrapy/selector copying scrapy/selector/unified.py -> build/lib/scrapy/selector copying scrapy/selector/__init__.py -> build/lib/scrapy/selector creating build/lib/scrapy/pipelines copying scrapy/pipelines/media.py -> build/lib/scrapy/pipelines copying scrapy/pipelines/images.py -> build/lib/scrapy/pipelines copying scrapy/pipelines/files.py -> build/lib/scrapy/pipelines copying scrapy/pipelines/__init__.py -> build/lib/scrapy/pipelines creating build/lib/scrapy/loader copying scrapy/loader/processors.py -> build/lib/scrapy/loader copying scrapy/loader/common.py -> build/lib/scrapy/loader copying scrapy/loader/__init__.py -> build/lib/scrapy/loader creating build/lib/scrapy/linkextractors copying scrapy/linkextractors/lxmlhtml.py -> build/lib/scrapy/linkextractors copying scrapy/linkextractors/__init__.py -> build/lib/scrapy/linkextractors creating build/lib/scrapy/http copying scrapy/http/headers.py -> build/lib/scrapy/http copying scrapy/http/cookies.py -> build/lib/scrapy/http copying scrapy/http/common.py -> build/lib/scrapy/http copying scrapy/http/__init__.py -> build/lib/scrapy/http creating build/lib/scrapy/extensions copying scrapy/extensions/throttle.py -> build/lib/scrapy/extensions copying scrapy/extensions/telnet.py -> build/lib/scrapy/extensions copying scrapy/extensions/statsmailer.py -> build/lib/scrapy/extensions copying scrapy/extensions/spiderstate.py -> build/lib/scrapy/extensions copying scrapy/extensions/postprocessing.py -> build/lib/scrapy/extensions copying scrapy/extensions/memusage.py -> build/lib/scrapy/extensions copying scrapy/extensions/memdebug.py -> build/lib/scrapy/extensions copying scrapy/extensions/logstats.py -> build/lib/scrapy/extensions copying scrapy/extensions/httpcache.py -> build/lib/scrapy/extensions copying scrapy/extensions/feedexport.py -> build/lib/scrapy/extensions copying scrapy/extensions/debug.py -> build/lib/scrapy/extensions copying scrapy/extensions/corestats.py -> build/lib/scrapy/extensions copying scrapy/extensions/closespider.py -> build/lib/scrapy/extensions copying scrapy/extensions/__init__.py -> build/lib/scrapy/extensions creating build/lib/scrapy/downloadermiddlewares copying scrapy/downloadermiddlewares/useragent.py -> build/lib/scrapy/downloadermiddlewares copying scrapy/downloadermiddlewares/stats.py -> build/lib/scrapy/downloadermiddlewares copying scrapy/downloadermiddlewares/robotstxt.py -> build/lib/scrapy/downloadermiddlewares copying scrapy/downloadermiddlewares/retry.py -> build/lib/scrapy/downloadermiddlewares copying scrapy/downloadermiddlewares/redirect.py -> build/lib/scrapy/downloadermiddlewares copying scrapy/downloadermiddlewares/httpproxy.py -> build/lib/scrapy/downloadermiddlewares copying scrapy/downloadermiddlewares/httpcompression.py -> build/lib/scrapy/downloadermiddlewares copying scrapy/downloadermiddlewares/httpcache.py -> build/lib/scrapy/downloadermiddlewares copying scrapy/downloadermiddlewares/httpauth.py -> build/lib/scrapy/downloadermiddlewares copying scrapy/downloadermiddlewares/downloadtimeout.py -> build/lib/scrapy/downloadermiddlewares copying scrapy/downloadermiddlewares/defaultheaders.py -> build/lib/scrapy/downloadermiddlewares copying scrapy/downloadermiddlewares/decompression.py -> build/lib/scrapy/downloadermiddlewares copying scrapy/downloadermiddlewares/cookies.py -> build/lib/scrapy/downloadermiddlewares copying scrapy/downloadermiddlewares/ajaxcrawl.py -> build/lib/scrapy/downloadermiddlewares copying scrapy/downloadermiddlewares/__init__.py -> build/lib/scrapy/downloadermiddlewares creating build/lib/scrapy/core copying scrapy/core/spidermw.py -> build/lib/scrapy/core copying scrapy/core/scraper.py -> build/lib/scrapy/core copying scrapy/core/scheduler.py -> build/lib/scrapy/core copying scrapy/core/engine.py -> build/lib/scrapy/core copying scrapy/core/__init__.py -> build/lib/scrapy/core creating build/lib/scrapy/contracts copying scrapy/contracts/default.py -> build/lib/scrapy/contracts copying scrapy/contracts/__init__.py -> build/lib/scrapy/contracts creating build/lib/scrapy/commands copying scrapy/commands/view.py -> build/lib/scrapy/commands copying scrapy/commands/version.py -> build/lib/scrapy/commands copying scrapy/commands/startproject.py -> build/lib/scrapy/commands copying scrapy/commands/shell.py -> build/lib/scrapy/commands copying scrapy/commands/settings.py -> build/lib/scrapy/commands copying scrapy/commands/runspider.py -> build/lib/scrapy/commands copying scrapy/commands/parse.py -> build/lib/scrapy/commands copying scrapy/commands/list.py -> build/lib/scrapy/commands copying scrapy/commands/genspider.py -> build/lib/scrapy/commands copying scrapy/commands/fetch.py -> build/lib/scrapy/commands copying scrapy/commands/edit.py -> build/lib/scrapy/commands copying scrapy/commands/crawl.py -> build/lib/scrapy/commands copying scrapy/commands/check.py -> build/lib/scrapy/commands copying scrapy/commands/bench.py -> build/lib/scrapy/commands copying scrapy/commands/__init__.py -> build/lib/scrapy/commands creating build/lib/scrapy/http/response copying scrapy/http/response/xml.py -> build/lib/scrapy/http/response copying scrapy/http/response/text.py -> build/lib/scrapy/http/response copying scrapy/http/response/html.py -> build/lib/scrapy/http/response copying scrapy/http/response/__init__.py -> build/lib/scrapy/http/response creating build/lib/scrapy/http/request copying scrapy/http/request/rpc.py -> build/lib/scrapy/http/request copying scrapy/http/request/json_request.py -> build/lib/scrapy/http/request copying scrapy/http/request/form.py -> build/lib/scrapy/http/request copying scrapy/http/request/__init__.py -> build/lib/scrapy/http/request creating build/lib/scrapy/core/http2 copying scrapy/core/http2/stream.py -> build/lib/scrapy/core/http2 copying scrapy/core/http2/protocol.py -> build/lib/scrapy/core/http2 copying scrapy/core/http2/agent.py -> build/lib/scrapy/core/http2 copying scrapy/core/http2/__init__.py -> build/lib/scrapy/core/http2 creating build/lib/scrapy/core/downloader copying scrapy/core/downloader/webclient.py -> build/lib/scrapy/core/downloader copying scrapy/core/downloader/tls.py -> build/lib/scrapy/core/downloader copying scrapy/core/downloader/middleware.py -> build/lib/scrapy/core/downloader copying scrapy/core/downloader/contextfactory.py -> build/lib/scrapy/core/downloader copying scrapy/core/downloader/__init__.py -> build/lib/scrapy/core/downloader creating build/lib/scrapy/core/downloader/handlers copying scrapy/core/downloader/handlers/s3.py -> build/lib/scrapy/core/downloader/handlers copying scrapy/core/downloader/handlers/http2.py -> build/lib/scrapy/core/downloader/handlers copying scrapy/core/downloader/handlers/http11.py -> build/lib/scrapy/core/downloader/handlers copying scrapy/core/downloader/handlers/http10.py -> build/lib/scrapy/core/downloader/handlers copying scrapy/core/downloader/handlers/http.py -> build/lib/scrapy/core/downloader/handlers copying scrapy/core/downloader/handlers/ftp.py -> build/lib/scrapy/core/downloader/handlers copying scrapy/core/downloader/handlers/file.py -> build/lib/scrapy/core/downloader/handlers copying scrapy/core/downloader/handlers/datauri.py -> build/lib/scrapy/core/downloader/handlers copying scrapy/core/downloader/handlers/__init__.py -> build/lib/scrapy/core/downloader/handlers running egg_info writing Scrapy.egg-info/PKG-INFO writing dependency_links to Scrapy.egg-info/dependency_links.txt writing entry points to Scrapy.egg-info/entry_points.txt writing requirements to Scrapy.egg-info/requires.txt writing top-level names to Scrapy.egg-info/top_level.txt reading manifest file 'Scrapy.egg-info/SOURCES.txt' reading manifest template 'MANIFEST.in' warning: no files found matching 'INSTALL' warning: no files found matching 'requirements-*.txt' warning: no files found matching 'license.txt' under directory 'scrapy' no previously-included directories found matching 'docs/build' warning: no files found matching '*' under directory 'bin' warning: no previously-included files matching '__pycache__' found anywhere in distribution warning: no previously-included files matching '*.py[cod]' found anywhere in distribution adding license file 'LICENSE' adding license file 'AUTHORS' writing manifest file 'Scrapy.egg-info/SOURCES.txt' /usr/lib/python3.9/site-packages/setuptools/command/build_py.py:201: _Warning: Package 'scrapy.templates.project' is absent from the `packages` configuration. !! ******************************************************************************** ############################ # Package would be ignored # ############################ Python recognizes 'scrapy.templates.project' as an importable package[^1], but it is absent from setuptools' `packages` configuration. This leads to an ambiguous overall configuration. If you want to distribute this package, please make sure that 'scrapy.templates.project' is explicitly added to the `packages` configuration field. Alternatively, you can also rely on setuptools' discovery methods (for example by using `find_namespace_packages(...)`/`find_namespace:` instead of `find_packages(...)`/`find:`). You can read more about "package discovery" on setuptools documentation page: - https://setuptools.pypa.io/en/latest/userguide/package_discovery.html If you don't want 'scrapy.templates.project' to be distributed and are already explicitly excluding 'scrapy.templates.project' via `find_namespace_packages(...)/find_namespace` or `find_packages(...)/find`, you can try to use `exclude_package_data`, or `include-package-data=False` in combination with a more fine grained `package-data` configuration. You can read more about "package data files" on setuptools documentation page: - https://setuptools.pypa.io/en/latest/userguide/datafiles.html [^1]: For Python, any directory (with suitable naming) can be imported, even if it does not contain any `.py` files. On the other hand, currently there is no concept of package data directory, all directories are treated like packages. ******************************************************************************** !! check.warn(importable) /usr/lib/python3.9/site-packages/setuptools/command/build_py.py:201: _Warning: Package 'scrapy.templates.project.module' is absent from the `packages` configuration. !! ******************************************************************************** ############################ # Package would be ignored # ############################ Python recognizes 'scrapy.templates.project.module' as an importable package[^1], but it is absent from setuptools' `packages` configuration. This leads to an ambiguous overall configuration. If you want to distribute this package, please make sure that 'scrapy.templates.project.module' is explicitly added to the `packages` configuration field. Alternatively, you can also rely on setuptools' discovery methods (for example by using `find_namespace_packages(...)`/`find_namespace:` instead of `find_packages(...)`/`find:`). You can read more about "package discovery" on setuptools documentation page: - https://setuptools.pypa.io/en/latest/userguide/package_discovery.html If you don't want 'scrapy.templates.project.module' to be distributed and are already explicitly excluding 'scrapy.templates.project.module' via `find_namespace_packages(...)/find_namespace` or `find_packages(...)/find`, you can try to use `exclude_package_data`, or `include-package-data=False` in combination with a more fine grained `package-data` configuration. You can read more about "package data files" on setuptools documentation page: - https://setuptools.pypa.io/en/latest/userguide/datafiles.html [^1]: For Python, any directory (with suitable naming) can be imported, even if it does not contain any `.py` files. On the other hand, currently there is no concept of package data directory, all directories are treated like packages. ******************************************************************************** !! check.warn(importable) /usr/lib/python3.9/site-packages/setuptools/command/build_py.py:201: _Warning: Package 'scrapy.templates.project.module.spiders' is absent from the `packages` configuration. !! ******************************************************************************** ############################ # Package would be ignored # ############################ Python recognizes 'scrapy.templates.project.module.spiders' as an importable package[^1], but it is absent from setuptools' `packages` configuration. This leads to an ambiguous overall configuration. If you want to distribute this package, please make sure that 'scrapy.templates.project.module.spiders' is explicitly added to the `packages` configuration field. Alternatively, you can also rely on setuptools' discovery methods (for example by using `find_namespace_packages(...)`/`find_namespace:` instead of `find_packages(...)`/`find:`). You can read more about "package discovery" on setuptools documentation page: - https://setuptools.pypa.io/en/latest/userguide/package_discovery.html If you don't want 'scrapy.templates.project.module.spiders' to be distributed and are already explicitly excluding 'scrapy.templates.project.module.spiders' via `find_namespace_packages(...)/find_namespace` or `find_packages(...)/find`, you can try to use `exclude_package_data`, or `include-package-data=False` in combination with a more fine grained `package-data` configuration. You can read more about "package data files" on setuptools documentation page: - https://setuptools.pypa.io/en/latest/userguide/datafiles.html [^1]: For Python, any directory (with suitable naming) can be imported, even if it does not contain any `.py` files. On the other hand, currently there is no concept of package data directory, all directories are treated like packages. ******************************************************************************** !! check.warn(importable) /usr/lib/python3.9/site-packages/setuptools/command/build_py.py:201: _Warning: Package 'scrapy.templates.spiders' is absent from the `packages` configuration. !! ******************************************************************************** ############################ # Package would be ignored # ############################ Python recognizes 'scrapy.templates.spiders' as an importable package[^1], but it is absent from setuptools' `packages` configuration. This leads to an ambiguous overall configuration. If you want to distribute this package, please make sure that 'scrapy.templates.spiders' is explicitly added to the `packages` configuration field. Alternatively, you can also rely on setuptools' discovery methods (for example by using `find_namespace_packages(...)`/`find_namespace:` instead of `find_packages(...)`/`find:`). You can read more about "package discovery" on setuptools documentation page: - https://setuptools.pypa.io/en/latest/userguide/package_discovery.html If you don't want 'scrapy.templates.spiders' to be distributed and are already explicitly excluding 'scrapy.templates.spiders' via `find_namespace_packages(...)/find_namespace` or `find_packages(...)/find`, you can try to use `exclude_package_data`, or `include-package-data=False` in combination with a more fine grained `package-data` configuration. You can read more about "package data files" on setuptools documentation page: - https://setuptools.pypa.io/en/latest/userguide/datafiles.html [^1]: For Python, any directory (with suitable naming) can be imported, even if it does not contain any `.py` files. On the other hand, currently there is no concept of package data directory, all directories are treated like packages. ******************************************************************************** !! check.warn(importable) copying scrapy/VERSION -> build/lib/scrapy copying scrapy/mime.types -> build/lib/scrapy creating build/lib/scrapy/templates creating build/lib/scrapy/templates/project copying scrapy/templates/project/scrapy.cfg -> build/lib/scrapy/templates/project creating build/lib/scrapy/templates/project/module copying scrapy/templates/project/module/__init__.py -> build/lib/scrapy/templates/project/module copying scrapy/templates/project/module/items.py.tmpl -> build/lib/scrapy/templates/project/module copying scrapy/templates/project/module/middlewares.py.tmpl -> build/lib/scrapy/templates/project/module copying scrapy/templates/project/module/pipelines.py.tmpl -> build/lib/scrapy/templates/project/module copying scrapy/templates/project/module/settings.py.tmpl -> build/lib/scrapy/templates/project/module creating build/lib/scrapy/templates/project/module/spiders copying scrapy/templates/project/module/spiders/__init__.py -> build/lib/scrapy/templates/project/module/spiders creating build/lib/scrapy/templates/spiders copying scrapy/templates/spiders/basic.tmpl -> build/lib/scrapy/templates/spiders copying scrapy/templates/spiders/crawl.tmpl -> build/lib/scrapy/templates/spiders copying scrapy/templates/spiders/csvfeed.tmpl -> build/lib/scrapy/templates/spiders copying scrapy/templates/spiders/xmlfeed.tmpl -> build/lib/scrapy/templates/spiders installing to build/bdist.linux-x86_64/wheel running install running install_lib creating build/bdist.linux-x86_64 creating build/bdist.linux-x86_64/wheel creating build/bdist.linux-x86_64/wheel/scrapy creating build/bdist.linux-x86_64/wheel/scrapy/templates creating build/bdist.linux-x86_64/wheel/scrapy/templates/spiders copying build/lib/scrapy/templates/spiders/xmlfeed.tmpl -> build/bdist.linux-x86_64/wheel/scrapy/templates/spiders copying build/lib/scrapy/templates/spiders/csvfeed.tmpl -> build/bdist.linux-x86_64/wheel/scrapy/templates/spiders copying build/lib/scrapy/templates/spiders/crawl.tmpl -> build/bdist.linux-x86_64/wheel/scrapy/templates/spiders copying build/lib/scrapy/templates/spiders/basic.tmpl -> build/bdist.linux-x86_64/wheel/scrapy/templates/spiders creating build/bdist.linux-x86_64/wheel/scrapy/templates/project creating build/bdist.linux-x86_64/wheel/scrapy/templates/project/module creating build/bdist.linux-x86_64/wheel/scrapy/templates/project/module/spiders copying build/lib/scrapy/templates/project/module/spiders/__init__.py -> build/bdist.linux-x86_64/wheel/scrapy/templates/project/module/spiders copying build/lib/scrapy/templates/project/module/settings.py.tmpl -> build/bdist.linux-x86_64/wheel/scrapy/templates/project/module copying build/lib/scrapy/templates/project/module/pipelines.py.tmpl -> build/bdist.linux-x86_64/wheel/scrapy/templates/project/module copying build/lib/scrapy/templates/project/module/middlewares.py.tmpl -> build/bdist.linux-x86_64/wheel/scrapy/templates/project/module copying build/lib/scrapy/templates/project/module/items.py.tmpl -> build/bdist.linux-x86_64/wheel/scrapy/templates/project/module copying build/lib/scrapy/templates/project/module/__init__.py -> build/bdist.linux-x86_64/wheel/scrapy/templates/project/module copying build/lib/scrapy/templates/project/scrapy.cfg -> build/bdist.linux-x86_64/wheel/scrapy/templates/project copying build/lib/scrapy/mime.types -> build/bdist.linux-x86_64/wheel/scrapy copying build/lib/scrapy/VERSION -> build/bdist.linux-x86_64/wheel/scrapy creating build/bdist.linux-x86_64/wheel/scrapy/commands copying build/lib/scrapy/commands/__init__.py -> build/bdist.linux-x86_64/wheel/scrapy/commands copying build/lib/scrapy/commands/bench.py -> build/bdist.linux-x86_64/wheel/scrapy/commands copying build/lib/scrapy/commands/check.py -> build/bdist.linux-x86_64/wheel/scrapy/commands copying build/lib/scrapy/commands/crawl.py -> build/bdist.linux-x86_64/wheel/scrapy/commands copying build/lib/scrapy/commands/edit.py -> build/bdist.linux-x86_64/wheel/scrapy/commands copying build/lib/scrapy/commands/fetch.py -> build/bdist.linux-x86_64/wheel/scrapy/commands copying build/lib/scrapy/commands/genspider.py -> build/bdist.linux-x86_64/wheel/scrapy/commands copying build/lib/scrapy/commands/list.py -> build/bdist.linux-x86_64/wheel/scrapy/commands copying build/lib/scrapy/commands/parse.py -> build/bdist.linux-x86_64/wheel/scrapy/commands copying build/lib/scrapy/commands/runspider.py -> build/bdist.linux-x86_64/wheel/scrapy/commands copying build/lib/scrapy/commands/settings.py -> build/bdist.linux-x86_64/wheel/scrapy/commands copying build/lib/scrapy/commands/shell.py -> build/bdist.linux-x86_64/wheel/scrapy/commands copying build/lib/scrapy/commands/startproject.py -> build/bdist.linux-x86_64/wheel/scrapy/commands copying build/lib/scrapy/commands/version.py -> build/bdist.linux-x86_64/wheel/scrapy/commands copying build/lib/scrapy/commands/view.py -> build/bdist.linux-x86_64/wheel/scrapy/commands creating build/bdist.linux-x86_64/wheel/scrapy/contracts copying build/lib/scrapy/contracts/__init__.py -> build/bdist.linux-x86_64/wheel/scrapy/contracts copying build/lib/scrapy/contracts/default.py -> build/bdist.linux-x86_64/wheel/scrapy/contracts creating build/bdist.linux-x86_64/wheel/scrapy/core creating build/bdist.linux-x86_64/wheel/scrapy/core/downloader creating build/bdist.linux-x86_64/wheel/scrapy/core/downloader/handlers copying build/lib/scrapy/core/downloader/handlers/__init__.py -> build/bdist.linux-x86_64/wheel/scrapy/core/downloader/handlers copying build/lib/scrapy/core/downloader/handlers/datauri.py -> build/bdist.linux-x86_64/wheel/scrapy/core/downloader/handlers copying build/lib/scrapy/core/downloader/handlers/file.py -> build/bdist.linux-x86_64/wheel/scrapy/core/downloader/handlers copying build/lib/scrapy/core/downloader/handlers/ftp.py -> build/bdist.linux-x86_64/wheel/scrapy/core/downloader/handlers copying build/lib/scrapy/core/downloader/handlers/http.py -> build/bdist.linux-x86_64/wheel/scrapy/core/downloader/handlers copying build/lib/scrapy/core/downloader/handlers/http10.py -> build/bdist.linux-x86_64/wheel/scrapy/core/downloader/handlers copying build/lib/scrapy/core/downloader/handlers/http11.py -> build/bdist.linux-x86_64/wheel/scrapy/core/downloader/handlers copying build/lib/scrapy/core/downloader/handlers/http2.py -> build/bdist.linux-x86_64/wheel/scrapy/core/downloader/handlers copying build/lib/scrapy/core/downloader/handlers/s3.py -> build/bdist.linux-x86_64/wheel/scrapy/core/downloader/handlers copying build/lib/scrapy/core/downloader/__init__.py -> build/bdist.linux-x86_64/wheel/scrapy/core/downloader copying build/lib/scrapy/core/downloader/contextfactory.py -> build/bdist.linux-x86_64/wheel/scrapy/core/downloader copying build/lib/scrapy/core/downloader/middleware.py -> build/bdist.linux-x86_64/wheel/scrapy/core/downloader copying build/lib/scrapy/core/downloader/tls.py -> build/bdist.linux-x86_64/wheel/scrapy/core/downloader copying build/lib/scrapy/core/downloader/webclient.py -> build/bdist.linux-x86_64/wheel/scrapy/core/downloader creating build/bdist.linux-x86_64/wheel/scrapy/core/http2 copying build/lib/scrapy/core/http2/__init__.py -> build/bdist.linux-x86_64/wheel/scrapy/core/http2 copying build/lib/scrapy/core/http2/agent.py -> build/bdist.linux-x86_64/wheel/scrapy/core/http2 copying build/lib/scrapy/core/http2/protocol.py -> build/bdist.linux-x86_64/wheel/scrapy/core/http2 copying build/lib/scrapy/core/http2/stream.py -> build/bdist.linux-x86_64/wheel/scrapy/core/http2 copying build/lib/scrapy/core/__init__.py -> build/bdist.linux-x86_64/wheel/scrapy/core copying build/lib/scrapy/core/engine.py -> build/bdist.linux-x86_64/wheel/scrapy/core copying build/lib/scrapy/core/scheduler.py -> build/bdist.linux-x86_64/wheel/scrapy/core copying build/lib/scrapy/core/scraper.py -> build/bdist.linux-x86_64/wheel/scrapy/core copying build/lib/scrapy/core/spidermw.py -> build/bdist.linux-x86_64/wheel/scrapy/core creating build/bdist.linux-x86_64/wheel/scrapy/downloadermiddlewares copying build/lib/scrapy/downloadermiddlewares/__init__.py -> build/bdist.linux-x86_64/wheel/scrapy/downloadermiddlewares copying build/lib/scrapy/downloadermiddlewares/ajaxcrawl.py -> build/bdist.linux-x86_64/wheel/scrapy/downloadermiddlewares copying build/lib/scrapy/downloadermiddlewares/cookies.py -> build/bdist.linux-x86_64/wheel/scrapy/downloadermiddlewares copying build/lib/scrapy/downloadermiddlewares/decompression.py -> build/bdist.linux-x86_64/wheel/scrapy/downloadermiddlewares copying build/lib/scrapy/downloadermiddlewares/defaultheaders.py -> build/bdist.linux-x86_64/wheel/scrapy/downloadermiddlewares copying build/lib/scrapy/downloadermiddlewares/downloadtimeout.py -> build/bdist.linux-x86_64/wheel/scrapy/downloadermiddlewares copying build/lib/scrapy/downloadermiddlewares/httpauth.py -> build/bdist.linux-x86_64/wheel/scrapy/downloadermiddlewares copying build/lib/scrapy/downloadermiddlewares/httpcache.py -> build/bdist.linux-x86_64/wheel/scrapy/downloadermiddlewares copying build/lib/scrapy/downloadermiddlewares/httpcompression.py -> build/bdist.linux-x86_64/wheel/scrapy/downloadermiddlewares copying build/lib/scrapy/downloadermiddlewares/httpproxy.py -> build/bdist.linux-x86_64/wheel/scrapy/downloadermiddlewares copying build/lib/scrapy/downloadermiddlewares/redirect.py -> build/bdist.linux-x86_64/wheel/scrapy/downloadermiddlewares copying build/lib/scrapy/downloadermiddlewares/retry.py -> build/bdist.linux-x86_64/wheel/scrapy/downloadermiddlewares copying build/lib/scrapy/downloadermiddlewares/robotstxt.py -> build/bdist.linux-x86_64/wheel/scrapy/downloadermiddlewares copying build/lib/scrapy/downloadermiddlewares/stats.py -> build/bdist.linux-x86_64/wheel/scrapy/downloadermiddlewares copying build/lib/scrapy/downloadermiddlewares/useragent.py -> build/bdist.linux-x86_64/wheel/scrapy/downloadermiddlewares creating build/bdist.linux-x86_64/wheel/scrapy/extensions copying build/lib/scrapy/extensions/__init__.py -> build/bdist.linux-x86_64/wheel/scrapy/extensions copying build/lib/scrapy/extensions/closespider.py -> build/bdist.linux-x86_64/wheel/scrapy/extensions copying build/lib/scrapy/extensions/corestats.py -> build/bdist.linux-x86_64/wheel/scrapy/extensions copying build/lib/scrapy/extensions/debug.py -> build/bdist.linux-x86_64/wheel/scrapy/extensions copying build/lib/scrapy/extensions/feedexport.py -> build/bdist.linux-x86_64/wheel/scrapy/extensions copying build/lib/scrapy/extensions/httpcache.py -> build/bdist.linux-x86_64/wheel/scrapy/extensions copying build/lib/scrapy/extensions/logstats.py -> build/bdist.linux-x86_64/wheel/scrapy/extensions copying build/lib/scrapy/extensions/memdebug.py -> build/bdist.linux-x86_64/wheel/scrapy/extensions copying build/lib/scrapy/extensions/memusage.py -> build/bdist.linux-x86_64/wheel/scrapy/extensions copying build/lib/scrapy/extensions/postprocessing.py -> build/bdist.linux-x86_64/wheel/scrapy/extensions copying build/lib/scrapy/extensions/spiderstate.py -> build/bdist.linux-x86_64/wheel/scrapy/extensions copying build/lib/scrapy/extensions/statsmailer.py -> build/bdist.linux-x86_64/wheel/scrapy/extensions copying build/lib/scrapy/extensions/telnet.py -> build/bdist.linux-x86_64/wheel/scrapy/extensions copying build/lib/scrapy/extensions/throttle.py -> build/bdist.linux-x86_64/wheel/scrapy/extensions creating build/bdist.linux-x86_64/wheel/scrapy/http creating build/bdist.linux-x86_64/wheel/scrapy/http/request copying build/lib/scrapy/http/request/__init__.py -> build/bdist.linux-x86_64/wheel/scrapy/http/request copying build/lib/scrapy/http/request/form.py -> build/bdist.linux-x86_64/wheel/scrapy/http/request copying build/lib/scrapy/http/request/json_request.py -> build/bdist.linux-x86_64/wheel/scrapy/http/request copying build/lib/scrapy/http/request/rpc.py -> build/bdist.linux-x86_64/wheel/scrapy/http/request creating build/bdist.linux-x86_64/wheel/scrapy/http/response copying build/lib/scrapy/http/response/__init__.py -> build/bdist.linux-x86_64/wheel/scrapy/http/response copying build/lib/scrapy/http/response/html.py -> build/bdist.linux-x86_64/wheel/scrapy/http/response copying build/lib/scrapy/http/response/text.py -> build/bdist.linux-x86_64/wheel/scrapy/http/response copying build/lib/scrapy/http/response/xml.py -> build/bdist.linux-x86_64/wheel/scrapy/http/response copying build/lib/scrapy/http/__init__.py -> build/bdist.linux-x86_64/wheel/scrapy/http copying build/lib/scrapy/http/common.py -> build/bdist.linux-x86_64/wheel/scrapy/http copying build/lib/scrapy/http/cookies.py -> build/bdist.linux-x86_64/wheel/scrapy/http copying build/lib/scrapy/http/headers.py -> build/bdist.linux-x86_64/wheel/scrapy/http creating build/bdist.linux-x86_64/wheel/scrapy/linkextractors copying build/lib/scrapy/linkextractors/__init__.py -> build/bdist.linux-x86_64/wheel/scrapy/linkextractors copying build/lib/scrapy/linkextractors/lxmlhtml.py -> build/bdist.linux-x86_64/wheel/scrapy/linkextractors creating build/bdist.linux-x86_64/wheel/scrapy/loader copying build/lib/scrapy/loader/__init__.py -> build/bdist.linux-x86_64/wheel/scrapy/loader copying build/lib/scrapy/loader/common.py -> build/bdist.linux-x86_64/wheel/scrapy/loader copying build/lib/scrapy/loader/processors.py -> build/bdist.linux-x86_64/wheel/scrapy/loader creating build/bdist.linux-x86_64/wheel/scrapy/pipelines copying build/lib/scrapy/pipelines/__init__.py -> build/bdist.linux-x86_64/wheel/scrapy/pipelines copying build/lib/scrapy/pipelines/files.py -> build/bdist.linux-x86_64/wheel/scrapy/pipelines copying build/lib/scrapy/pipelines/images.py -> build/bdist.linux-x86_64/wheel/scrapy/pipelines copying build/lib/scrapy/pipelines/media.py -> build/bdist.linux-x86_64/wheel/scrapy/pipelines creating build/bdist.linux-x86_64/wheel/scrapy/selector copying build/lib/scrapy/selector/__init__.py -> build/bdist.linux-x86_64/wheel/scrapy/selector copying build/lib/scrapy/selector/unified.py -> build/bdist.linux-x86_64/wheel/scrapy/selector creating build/bdist.linux-x86_64/wheel/scrapy/settings copying build/lib/scrapy/settings/__init__.py -> build/bdist.linux-x86_64/wheel/scrapy/settings copying build/lib/scrapy/settings/default_settings.py -> build/bdist.linux-x86_64/wheel/scrapy/settings creating build/bdist.linux-x86_64/wheel/scrapy/spidermiddlewares copying build/lib/scrapy/spidermiddlewares/__init__.py -> build/bdist.linux-x86_64/wheel/scrapy/spidermiddlewares copying build/lib/scrapy/spidermiddlewares/depth.py -> build/bdist.linux-x86_64/wheel/scrapy/spidermiddlewares copying build/lib/scrapy/spidermiddlewares/httperror.py -> build/bdist.linux-x86_64/wheel/scrapy/spidermiddlewares copying build/lib/scrapy/spidermiddlewares/offsite.py -> build/bdist.linux-x86_64/wheel/scrapy/spidermiddlewares copying build/lib/scrapy/spidermiddlewares/referer.py -> build/bdist.linux-x86_64/wheel/scrapy/spidermiddlewares copying build/lib/scrapy/spidermiddlewares/urllength.py -> build/bdist.linux-x86_64/wheel/scrapy/spidermiddlewares creating build/bdist.linux-x86_64/wheel/scrapy/spiders copying build/lib/scrapy/spiders/__init__.py -> build/bdist.linux-x86_64/wheel/scrapy/spiders copying build/lib/scrapy/spiders/crawl.py -> build/bdist.linux-x86_64/wheel/scrapy/spiders copying build/lib/scrapy/spiders/feed.py -> build/bdist.linux-x86_64/wheel/scrapy/spiders copying build/lib/scrapy/spiders/init.py -> build/bdist.linux-x86_64/wheel/scrapy/spiders copying build/lib/scrapy/spiders/sitemap.py -> build/bdist.linux-x86_64/wheel/scrapy/spiders creating build/bdist.linux-x86_64/wheel/scrapy/utils copying build/lib/scrapy/utils/__init__.py -> build/bdist.linux-x86_64/wheel/scrapy/utils copying build/lib/scrapy/utils/asyncgen.py -> build/bdist.linux-x86_64/wheel/scrapy/utils copying build/lib/scrapy/utils/benchserver.py -> build/bdist.linux-x86_64/wheel/scrapy/utils copying build/lib/scrapy/utils/boto.py -> build/bdist.linux-x86_64/wheel/scrapy/utils copying build/lib/scrapy/utils/conf.py -> build/bdist.linux-x86_64/wheel/scrapy/utils copying build/lib/scrapy/utils/console.py -> build/bdist.linux-x86_64/wheel/scrapy/utils copying build/lib/scrapy/utils/curl.py -> build/bdist.linux-x86_64/wheel/scrapy/utils copying build/lib/scrapy/utils/datatypes.py -> build/bdist.linux-x86_64/wheel/scrapy/utils copying build/lib/scrapy/utils/decorators.py -> build/bdist.linux-x86_64/wheel/scrapy/utils copying build/lib/scrapy/utils/defer.py -> build/bdist.linux-x86_64/wheel/scrapy/utils copying build/lib/scrapy/utils/deprecate.py -> build/bdist.linux-x86_64/wheel/scrapy/utils copying build/lib/scrapy/utils/display.py -> build/bdist.linux-x86_64/wheel/scrapy/utils copying build/lib/scrapy/utils/engine.py -> build/bdist.linux-x86_64/wheel/scrapy/utils copying build/lib/scrapy/utils/ftp.py -> build/bdist.linux-x86_64/wheel/scrapy/utils copying build/lib/scrapy/utils/gz.py -> build/bdist.linux-x86_64/wheel/scrapy/utils copying build/lib/scrapy/utils/httpobj.py -> build/bdist.linux-x86_64/wheel/scrapy/utils copying build/lib/scrapy/utils/iterators.py -> build/bdist.linux-x86_64/wheel/scrapy/utils copying build/lib/scrapy/utils/job.py -> build/bdist.linux-x86_64/wheel/scrapy/utils copying build/lib/scrapy/utils/log.py -> build/bdist.linux-x86_64/wheel/scrapy/utils copying build/lib/scrapy/utils/misc.py -> build/bdist.linux-x86_64/wheel/scrapy/utils copying build/lib/scrapy/utils/ossignal.py -> build/bdist.linux-x86_64/wheel/scrapy/utils copying build/lib/scrapy/utils/project.py -> build/bdist.linux-x86_64/wheel/scrapy/utils copying build/lib/scrapy/utils/python.py -> build/bdist.linux-x86_64/wheel/scrapy/utils copying build/lib/scrapy/utils/reactor.py -> build/bdist.linux-x86_64/wheel/scrapy/utils copying build/lib/scrapy/utils/reqser.py -> build/bdist.linux-x86_64/wheel/scrapy/utils copying build/lib/scrapy/utils/request.py -> build/bdist.linux-x86_64/wheel/scrapy/utils copying build/lib/scrapy/utils/response.py -> build/bdist.linux-x86_64/wheel/scrapy/utils copying build/lib/scrapy/utils/serialize.py -> build/bdist.linux-x86_64/wheel/scrapy/utils copying build/lib/scrapy/utils/signal.py -> build/bdist.linux-x86_64/wheel/scrapy/utils copying build/lib/scrapy/utils/sitemap.py -> build/bdist.linux-x86_64/wheel/scrapy/utils copying build/lib/scrapy/utils/spider.py -> build/bdist.linux-x86_64/wheel/scrapy/utils copying build/lib/scrapy/utils/ssl.py -> build/bdist.linux-x86_64/wheel/scrapy/utils copying build/lib/scrapy/utils/template.py -> build/bdist.linux-x86_64/wheel/scrapy/utils copying build/lib/scrapy/utils/test.py -> build/bdist.linux-x86_64/wheel/scrapy/utils copying build/lib/scrapy/utils/testproc.py -> build/bdist.linux-x86_64/wheel/scrapy/utils copying build/lib/scrapy/utils/testsite.py -> build/bdist.linux-x86_64/wheel/scrapy/utils copying build/lib/scrapy/utils/trackref.py -> build/bdist.linux-x86_64/wheel/scrapy/utils copying build/lib/scrapy/utils/url.py -> build/bdist.linux-x86_64/wheel/scrapy/utils copying build/lib/scrapy/utils/versions.py -> build/bdist.linux-x86_64/wheel/scrapy/utils copying build/lib/scrapy/__init__.py -> build/bdist.linux-x86_64/wheel/scrapy copying build/lib/scrapy/__main__.py -> build/bdist.linux-x86_64/wheel/scrapy copying build/lib/scrapy/cmdline.py -> build/bdist.linux-x86_64/wheel/scrapy copying build/lib/scrapy/crawler.py -> build/bdist.linux-x86_64/wheel/scrapy copying build/lib/scrapy/dupefilters.py -> build/bdist.linux-x86_64/wheel/scrapy copying build/lib/scrapy/exceptions.py -> build/bdist.linux-x86_64/wheel/scrapy copying build/lib/scrapy/exporters.py -> build/bdist.linux-x86_64/wheel/scrapy copying build/lib/scrapy/extension.py -> build/bdist.linux-x86_64/wheel/scrapy copying build/lib/scrapy/interfaces.py -> build/bdist.linux-x86_64/wheel/scrapy copying build/lib/scrapy/item.py -> build/bdist.linux-x86_64/wheel/scrapy copying build/lib/scrapy/link.py -> build/bdist.linux-x86_64/wheel/scrapy copying build/lib/scrapy/logformatter.py -> build/bdist.linux-x86_64/wheel/scrapy copying build/lib/scrapy/mail.py -> build/bdist.linux-x86_64/wheel/scrapy copying build/lib/scrapy/middleware.py -> build/bdist.linux-x86_64/wheel/scrapy copying build/lib/scrapy/pqueues.py -> build/bdist.linux-x86_64/wheel/scrapy copying build/lib/scrapy/resolver.py -> build/bdist.linux-x86_64/wheel/scrapy copying build/lib/scrapy/responsetypes.py -> build/bdist.linux-x86_64/wheel/scrapy copying build/lib/scrapy/robotstxt.py -> build/bdist.linux-x86_64/wheel/scrapy copying build/lib/scrapy/shell.py -> build/bdist.linux-x86_64/wheel/scrapy copying build/lib/scrapy/signalmanager.py -> build/bdist.linux-x86_64/wheel/scrapy copying build/lib/scrapy/signals.py -> build/bdist.linux-x86_64/wheel/scrapy copying build/lib/scrapy/spiderloader.py -> build/bdist.linux-x86_64/wheel/scrapy copying build/lib/scrapy/squeues.py -> build/bdist.linux-x86_64/wheel/scrapy copying build/lib/scrapy/statscollectors.py -> build/bdist.linux-x86_64/wheel/scrapy running install_egg_info Copying Scrapy.egg-info to build/bdist.linux-x86_64/wheel/Scrapy-2.9.0-py3.9.egg-info running install_scripts adding license file "LICENSE" (matched pattern "LICEN[CS]E*") adding license file "AUTHORS" (matched pattern "AUTHORS*") creating build/bdist.linux-x86_64/wheel/Scrapy-2.9.0.dist-info/WHEEL creating '/builddir/build/BUILD/Scrapy-2.9.0/.pyproject-builddir/pip-wheel-z97zb_jo/.tmp-zxd3cvj3/Scrapy-2.9.0-py2.py3-none-any.whl' and adding 'build/bdist.linux-x86_64/wheel' to it adding 'scrapy/VERSION' adding 'scrapy/__init__.py' adding 'scrapy/__main__.py' adding 'scrapy/cmdline.py' adding 'scrapy/crawler.py' adding 'scrapy/dupefilters.py' adding 'scrapy/exceptions.py' adding 'scrapy/exporters.py' adding 'scrapy/extension.py' adding 'scrapy/interfaces.py' adding 'scrapy/item.py' adding 'scrapy/link.py' adding 'scrapy/logformatter.py' adding 'scrapy/mail.py' adding 'scrapy/middleware.py' adding 'scrapy/mime.types' adding 'scrapy/pqueues.py' adding 'scrapy/resolver.py' adding 'scrapy/responsetypes.py' adding 'scrapy/robotstxt.py' adding 'scrapy/shell.py' adding 'scrapy/signalmanager.py' adding 'scrapy/signals.py' adding 'scrapy/spiderloader.py' adding 'scrapy/squeues.py' adding 'scrapy/statscollectors.py' adding 'scrapy/commands/__init__.py' adding 'scrapy/commands/bench.py' adding 'scrapy/commands/check.py' adding 'scrapy/commands/crawl.py' adding 'scrapy/commands/edit.py' adding 'scrapy/commands/fetch.py' adding 'scrapy/commands/genspider.py' adding 'scrapy/commands/list.py' adding 'scrapy/commands/parse.py' adding 'scrapy/commands/runspider.py' adding 'scrapy/commands/settings.py' adding 'scrapy/commands/shell.py' adding 'scrapy/commands/startproject.py' adding 'scrapy/commands/version.py' adding 'scrapy/commands/view.py' adding 'scrapy/contracts/__init__.py' adding 'scrapy/contracts/default.py' adding 'scrapy/core/__init__.py' adding 'scrapy/core/engine.py' adding 'scrapy/core/scheduler.py' adding 'scrapy/core/scraper.py' adding 'scrapy/core/spidermw.py' adding 'scrapy/core/downloader/__init__.py' adding 'scrapy/core/downloader/contextfactory.py' adding 'scrapy/core/downloader/middleware.py' adding 'scrapy/core/downloader/tls.py' adding 'scrapy/core/downloader/webclient.py' adding 'scrapy/core/downloader/handlers/__init__.py' adding 'scrapy/core/downloader/handlers/datauri.py' adding 'scrapy/core/downloader/handlers/file.py' adding 'scrapy/core/downloader/handlers/ftp.py' adding 'scrapy/core/downloader/handlers/http.py' adding 'scrapy/core/downloader/handlers/http10.py' adding 'scrapy/core/downloader/handlers/http11.py' adding 'scrapy/core/downloader/handlers/http2.py' adding 'scrapy/core/downloader/handlers/s3.py' adding 'scrapy/core/http2/__init__.py' adding 'scrapy/core/http2/agent.py' adding 'scrapy/core/http2/protocol.py' adding 'scrapy/core/http2/stream.py' adding 'scrapy/downloadermiddlewares/__init__.py' adding 'scrapy/downloadermiddlewares/ajaxcrawl.py' adding 'scrapy/downloadermiddlewares/cookies.py' adding 'scrapy/downloadermiddlewares/decompression.py' adding 'scrapy/downloadermiddlewares/defaultheaders.py' adding 'scrapy/downloadermiddlewares/downloadtimeout.py' adding 'scrapy/downloadermiddlewares/httpauth.py' adding 'scrapy/downloadermiddlewares/httpcache.py' adding 'scrapy/downloadermiddlewares/httpcompression.py' adding 'scrapy/downloadermiddlewares/httpproxy.py' adding 'scrapy/downloadermiddlewares/redirect.py' adding 'scrapy/downloadermiddlewares/retry.py' adding 'scrapy/downloadermiddlewares/robotstxt.py' adding 'scrapy/downloadermiddlewares/stats.py' adding 'scrapy/downloadermiddlewares/useragent.py' adding 'scrapy/extensions/__init__.py' adding 'scrapy/extensions/closespider.py' adding 'scrapy/extensions/corestats.py' adding 'scrapy/extensions/debug.py' adding 'scrapy/extensions/feedexport.py' adding 'scrapy/extensions/httpcache.py' adding 'scrapy/extensions/logstats.py' adding 'scrapy/extensions/memdebug.py' adding 'scrapy/extensions/memusage.py' adding 'scrapy/extensions/postprocessing.py' adding 'scrapy/extensions/spiderstate.py' adding 'scrapy/extensions/statsmailer.py' adding 'scrapy/extensions/telnet.py' adding 'scrapy/extensions/throttle.py' adding 'scrapy/http/__init__.py' adding 'scrapy/http/common.py' adding 'scrapy/http/cookies.py' adding 'scrapy/http/headers.py' adding 'scrapy/http/request/__init__.py' adding 'scrapy/http/request/form.py' adding 'scrapy/http/request/json_request.py' adding 'scrapy/http/request/rpc.py' adding 'scrapy/http/response/__init__.py' adding 'scrapy/http/response/html.py' adding 'scrapy/http/response/text.py' adding 'scrapy/http/response/xml.py' adding 'scrapy/linkextractors/__init__.py' adding 'scrapy/linkextractors/lxmlhtml.py' adding 'scrapy/loader/__init__.py' adding 'scrapy/loader/common.py' adding 'scrapy/loader/processors.py' adding 'scrapy/pipelines/__init__.py' adding 'scrapy/pipelines/files.py' adding 'scrapy/pipelines/images.py' adding 'scrapy/pipelines/media.py' adding 'scrapy/selector/__init__.py' adding 'scrapy/selector/unified.py' adding 'scrapy/settings/__init__.py' adding 'scrapy/settings/default_settings.py' adding 'scrapy/spidermiddlewares/__init__.py' adding 'scrapy/spidermiddlewares/depth.py' adding 'scrapy/spidermiddlewares/httperror.py' adding 'scrapy/spidermiddlewares/offsite.py' adding 'scrapy/spidermiddlewares/referer.py' adding 'scrapy/spidermiddlewares/urllength.py' adding 'scrapy/spiders/__init__.py' adding 'scrapy/spiders/crawl.py' adding 'scrapy/spiders/feed.py' adding 'scrapy/spiders/init.py' adding 'scrapy/spiders/sitemap.py' adding 'scrapy/templates/project/scrapy.cfg' adding 'scrapy/templates/project/module/__init__.py' adding 'scrapy/templates/project/module/items.py.tmpl' adding 'scrapy/templates/project/module/middlewares.py.tmpl' adding 'scrapy/templates/project/module/pipelines.py.tmpl' adding 'scrapy/templates/project/module/settings.py.tmpl' adding 'scrapy/templates/project/module/spiders/__init__.py' adding 'scrapy/templates/spiders/basic.tmpl' adding 'scrapy/templates/spiders/crawl.tmpl' adding 'scrapy/templates/spiders/csvfeed.tmpl' adding 'scrapy/templates/spiders/xmlfeed.tmpl' adding 'scrapy/utils/__init__.py' adding 'scrapy/utils/asyncgen.py' adding 'scrapy/utils/benchserver.py' adding 'scrapy/utils/boto.py' adding 'scrapy/utils/conf.py' adding 'scrapy/utils/console.py' adding 'scrapy/utils/curl.py' adding 'scrapy/utils/datatypes.py' adding 'scrapy/utils/decorators.py' adding 'scrapy/utils/defer.py' adding 'scrapy/utils/deprecate.py' adding 'scrapy/utils/display.py' adding 'scrapy/utils/engine.py' adding 'scrapy/utils/ftp.py' adding 'scrapy/utils/gz.py' adding 'scrapy/utils/httpobj.py' adding 'scrapy/utils/iterators.py' adding 'scrapy/utils/job.py' adding 'scrapy/utils/log.py' adding 'scrapy/utils/misc.py' adding 'scrapy/utils/ossignal.py' adding 'scrapy/utils/project.py' adding 'scrapy/utils/python.py' adding 'scrapy/utils/reactor.py' adding 'scrapy/utils/reqser.py' adding 'scrapy/utils/request.py' adding 'scrapy/utils/response.py' adding 'scrapy/utils/serialize.py' adding 'scrapy/utils/signal.py' adding 'scrapy/utils/sitemap.py' adding 'scrapy/utils/spider.py' adding 'scrapy/utils/ssl.py' adding 'scrapy/utils/template.py' adding 'scrapy/utils/test.py' adding 'scrapy/utils/testproc.py' adding 'scrapy/utils/testsite.py' adding 'scrapy/utils/trackref.py' adding 'scrapy/utils/url.py' adding 'scrapy/utils/versions.py' adding 'Scrapy-2.9.0.dist-info/AUTHORS' adding 'Scrapy-2.9.0.dist-info/LICENSE' adding 'Scrapy-2.9.0.dist-info/METADATA' adding 'Scrapy-2.9.0.dist-info/WHEEL' adding 'Scrapy-2.9.0.dist-info/entry_points.txt' adding 'Scrapy-2.9.0.dist-info/top_level.txt' adding 'Scrapy-2.9.0.dist-info/RECORD' removing build/bdist.linux-x86_64/wheel Building wheel for Scrapy (pyproject.toml): finished with status 'done' Created wheel for Scrapy: filename=Scrapy-2.9.0-py2.py3-none-any.whl size=277179 sha256=6f99abe01f2c6bfd4db9b49dcc50c23cf09eca65a11e7b913f79a0b31a803475 Stored in directory: /builddir/.cache/pip/wheels/86/74/36/6ac49ac5082a71f873f4eedaba0395ec08e53ba2388aff5efa Successfully built Scrapy + RPM_EC=0 ++ jobs -p + exit 0 Executing(%install): /bin/sh -e /var/tmp/rpm-tmp.CUrZTL + umask 022 + cd /builddir/build/BUILD + '[' /builddir/build/BUILDROOT/python-scrapy-2.9.0-1.el9.x86_64 '!=' / ']' + rm -rf /builddir/build/BUILDROOT/python-scrapy-2.9.0-1.el9.x86_64 ++ dirname /builddir/build/BUILDROOT/python-scrapy-2.9.0-1.el9.x86_64 + mkdir -p /builddir/build/BUILDROOT + mkdir /builddir/build/BUILDROOT/python-scrapy-2.9.0-1.el9.x86_64 + cd Scrapy-2.9.0 ++ ls /builddir/build/BUILD/Scrapy-2.9.0/pyproject-wheeldir/Scrapy-2.9.0-py2.py3-none-any.whl ++ xargs basename --multiple ++ sed -E 's/([^-]+)-([^-]+)-.+\.whl/\1==\2/' + specifier=Scrapy==2.9.0 + TMPDIR=/builddir/build/BUILD/Scrapy-2.9.0/.pyproject-builddir + /usr/bin/python3 -m pip install --root /builddir/build/BUILDROOT/python-scrapy-2.9.0-1.el9.x86_64 --prefix /usr --no-deps --disable-pip-version-check --progress-bar off --verbose --ignore-installed --no-warn-script-location --no-index --no-cache-dir --find-links /builddir/build/BUILD/Scrapy-2.9.0/pyproject-wheeldir Scrapy==2.9.0 Using pip 22.0.4 from /usr/lib/python3.9/site-packages/pip (python 3.9) Looking in links: /builddir/build/BUILD/Scrapy-2.9.0/pyproject-wheeldir Processing ./pyproject-wheeldir/Scrapy-2.9.0-py2.py3-none-any.whl Installing collected packages: Scrapy Creating /builddir/build/BUILDROOT/python-scrapy-2.9.0-1.el9.x86_64/usr/bin changing mode of /builddir/build/BUILDROOT/python-scrapy-2.9.0-1.el9.x86_64/usr/bin/scrapy to 755 Successfully installed Scrapy-2.9.0 + '[' -d /builddir/build/BUILDROOT/python-scrapy-2.9.0-1.el9.x86_64/usr/bin ']' + '[' -f /usr/bin/pathfix3.9.py ']' + pathfix=/usr/bin/pathfix3.9.py + '[' -z s ']' + shebang_flags=-kas + /usr/bin/pathfix3.9.py -pni /usr/bin/python3 -kas /builddir/build/BUILDROOT/python-scrapy-2.9.0-1.el9.x86_64/usr/bin/scrapy /builddir/build/BUILDROOT/python-scrapy-2.9.0-1.el9.x86_64/usr/bin/scrapy: updating + rm -rfv /builddir/build/BUILDROOT/python-scrapy-2.9.0-1.el9.x86_64/usr/bin/__pycache__ + rm -f /builddir/build/BUILD/python-scrapy-2.9.0-1.el9.x86_64-pyproject-ghost-distinfo + site_dirs=() + '[' -d /builddir/build/BUILDROOT/python-scrapy-2.9.0-1.el9.x86_64/usr/lib/python3.9/site-packages ']' + site_dirs+=("/usr/lib/python3.9/site-packages") + '[' /builddir/build/BUILDROOT/python-scrapy-2.9.0-1.el9.x86_64/usr/lib64/python3.9/site-packages '!=' /builddir/build/BUILDROOT/python-scrapy-2.9.0-1.el9.x86_64/usr/lib/python3.9/site-packages ']' + '[' -d /builddir/build/BUILDROOT/python-scrapy-2.9.0-1.el9.x86_64/usr/lib64/python3.9/site-packages ']' + for site_dir in ${site_dirs[@]} + for distinfo in /builddir/build/BUILDROOT/python-scrapy-2.9.0-1.el9.x86_64$site_dir/*.dist-info + echo '%ghost /usr/lib/python3.9/site-packages/Scrapy-2.9.0.dist-info' + sed -i s/pip/rpm/ /builddir/build/BUILDROOT/python-scrapy-2.9.0-1.el9.x86_64/usr/lib/python3.9/site-packages/Scrapy-2.9.0.dist-info/INSTALLER + PYTHONPATH=/usr/lib/rpm/redhat + /usr/bin/python3 -B /usr/lib/rpm/redhat/pyproject_preprocess_record.py --buildroot /builddir/build/BUILDROOT/python-scrapy-2.9.0-1.el9.x86_64 --record /builddir/build/BUILDROOT/python-scrapy-2.9.0-1.el9.x86_64/usr/lib/python3.9/site-packages/Scrapy-2.9.0.dist-info/RECORD --output /builddir/build/BUILD/python-scrapy-2.9.0-1.el9.x86_64-pyproject-record + rm -fv /builddir/build/BUILDROOT/python-scrapy-2.9.0-1.el9.x86_64/usr/lib/python3.9/site-packages/Scrapy-2.9.0.dist-info/RECORD removed '/builddir/build/BUILDROOT/python-scrapy-2.9.0-1.el9.x86_64/usr/lib/python3.9/site-packages/Scrapy-2.9.0.dist-info/RECORD' + rm -fv /builddir/build/BUILDROOT/python-scrapy-2.9.0-1.el9.x86_64/usr/lib/python3.9/site-packages/Scrapy-2.9.0.dist-info/REQUESTED removed '/builddir/build/BUILDROOT/python-scrapy-2.9.0-1.el9.x86_64/usr/lib/python3.9/site-packages/Scrapy-2.9.0.dist-info/REQUESTED' ++ wc -l /builddir/build/BUILD/python-scrapy-2.9.0-1.el9.x86_64-pyproject-ghost-distinfo ++ cut -f1 '-d ' + lines=1 + '[' 1 -ne 1 ']' + /usr/bin/python3 /usr/lib/rpm/redhat/pyproject_save_files.py --output-files /builddir/build/BUILD/python-scrapy-2.9.0-1.el9.x86_64-pyproject-files --output-modules /builddir/build/BUILD/python-scrapy-2.9.0-1.el9.x86_64-pyproject-modules --buildroot /builddir/build/BUILDROOT/python-scrapy-2.9.0-1.el9.x86_64 --sitelib /usr/lib/python3.9/site-packages --sitearch /usr/lib64/python3.9/site-packages --python-version 3.9 --pyproject-record /builddir/build/BUILD/python-scrapy-2.9.0-1.el9.x86_64-pyproject-record --prefix /usr '*' +auto + /usr/lib/rpm/find-debuginfo.sh -j2 --strict-build-id -m -i --build-id-seed 2.9.0-1.el9 --unique-debug-suffix -2.9.0-1.el9.x86_64 --unique-debug-src-base python-scrapy-2.9.0-1.el9.x86_64 --run-dwz --dwz-low-mem-die-limit 10000000 --dwz-max-die-limit 110000000 --remove-section .gnu.build.attributes -S debugsourcefiles.list /builddir/build/BUILD/Scrapy-2.9.0 find: 'debug': No such file or directory + /usr/lib/rpm/check-buildroot + /usr/lib/rpm/redhat/brp-ldconfig + /usr/lib/rpm/brp-compress + /usr/lib/rpm/redhat/brp-strip-lto /usr/bin/strip + /usr/lib/rpm/brp-strip-static-archive /usr/bin/strip + /usr/lib/rpm/redhat/brp-python-bytecompile '' 1 0 Bytecompiling .py files below /builddir/build/BUILDROOT/python-scrapy-2.9.0-1.el9.x86_64/usr/lib/python3.9 using python3.9 + /usr/lib/rpm/brp-python-hardlink + /usr/lib/rpm/redhat/brp-mangle-shebangs Executing(%check): /bin/sh -e /var/tmp/rpm-tmp.B2ApMN + umask 022 + cd /builddir/build/BUILD + cd Scrapy-2.9.0 + '[' '!' -f /builddir/build/BUILD/python-scrapy-2.9.0-1.el9.x86_64-pyproject-modules ']' + PATH=/builddir/build/BUILDROOT/python-scrapy-2.9.0-1.el9.x86_64/usr/bin:/builddir/.local/bin:/builddir/bin:/usr/bin:/bin:/usr/sbin:/sbin:/usr/local/sbin + PYTHONPATH=/builddir/build/BUILDROOT/python-scrapy-2.9.0-1.el9.x86_64/usr/lib64/python3.9/site-packages:/builddir/build/BUILDROOT/python-scrapy-2.9.0-1.el9.x86_64/usr/lib/python3.9/site-packages + _PYTHONSITE=/builddir/build/BUILDROOT/python-scrapy-2.9.0-1.el9.x86_64/usr/lib64/python3.9/site-packages:/builddir/build/BUILDROOT/python-scrapy-2.9.0-1.el9.x86_64/usr/lib/python3.9/site-packages + PYTHONDONTWRITEBYTECODE=1 + /usr/bin/python3 -s /usr/lib/rpm/redhat/import_all_modules.py -f /builddir/build/BUILD/python-scrapy-2.9.0-1.el9.x86_64-pyproject-modules -t Check import: scrapy + RPM_EC=0 ++ jobs -p + exit 0 Processing files: python3-scrapy-2.9.0-1.el9.noarch Provides: python-scrapy = 2.9.0-1.el9 python3-scrapy = 2.9.0-1.el9 python3.9-scrapy = 2.9.0-1.el9 python3.9dist(scrapy) = 2.9 python3dist(scrapy) = 2.9 Requires(rpmlib): rpmlib(CompressedFileNames) <= 3.0.4-1 rpmlib(FileDigests) <= 4.6.0-1 rpmlib(PartialHardlinkSets) <= 4.0.4-1 rpmlib(PayloadFilesHavePrefix) <= 4.0-1 Requires: /usr/bin/python3 python(abi) = 3.9 python3.9dist(cryptography) >= 3.4.6 python3.9dist(cssselect) >= 0.9.1 python3.9dist(itemadapter) >= 0.1 python3.9dist(itemloaders) >= 1.0.1 python3.9dist(lxml) >= 4.3 python3.9dist(packaging) python3.9dist(parsel) >= 1.5 python3.9dist(protego) >= 0.1.15 python3.9dist(pydispatcher) >= 2.0.5 python3.9dist(pyopenssl) >= 21 python3.9dist(queuelib) >= 1.4.2 python3.9dist(service-identity) >= 18.1 python3.9dist(setuptools) python3.9dist(tldextract) python3.9dist(twisted) >= 18.9 python3.9dist(w3lib) >= 1.17 python3.9dist(zope-interface) >= 5.1 Obsoletes: python39-scrapy < 2.9.0-1.el9 Checking for unpackaged file(s): /usr/lib/rpm/check-files /builddir/build/BUILDROOT/python-scrapy-2.9.0-1.el9.x86_64 Wrote: /builddir/build/SRPMS/python-scrapy-2.9.0-1.el9.src.rpm Wrote: /builddir/build/RPMS/python3-scrapy-2.9.0-1.el9.noarch.rpm Executing(%clean): /bin/sh -e /var/tmp/rpm-tmp.dWYxoR + umask 022 + cd /builddir/build/BUILD + cd Scrapy-2.9.0 + /usr/bin/rm -rf /builddir/build/BUILDROOT/python-scrapy-2.9.0-1.el9.x86_64 + RPM_EC=0 ++ jobs -p + exit 0 Child return code was: 0