Mock Version: 3.0 ENTER ['do_with_status'](['bash', '--login', '-c', '/usr/bin/rpmbuild -bs --target x86_64 --nodeps /builddir/build/SPECS/python-scrapy-dynamic-spiders.spec'], chrootPath='/var/lib/mock/fedora-rawhide-x86_64-1649687776.743359/root'env={'TERM': 'vt100', 'SHELL': '/bin/bash', 'HOME': '/builddir', 'HOSTNAME': 'mock', 'PATH': '/usr/bin:/bin:/usr/sbin:/sbin', 'PROMPT_COMMAND': 'printf "\\033]0;\\007"', 'PS1': ' \\s-\\v\\$ ', 'LANG': 'C.UTF-8'}shell=Falselogger=timeout=0uid=1001gid=135user='mockbuild'nspawn_args=['--capability=cap_ipc_lock', '--rlimit=RLIMIT_NOFILE=10240', '--capability=cap_ipc_lock', '--bind=/tmp/mock-resolv.ym2m_f69:/etc/resolv.conf', '--bind=/dev/btrfs-control', '--bind=/dev/loop-control', '--bind=/dev/loop0', '--bind=/dev/loop1', '--bind=/dev/loop2', '--bind=/dev/loop3', '--bind=/dev/loop4', '--bind=/dev/loop5', '--bind=/dev/loop6', '--bind=/dev/loop7', '--bind=/dev/loop8', '--bind=/dev/loop9', '--bind=/dev/loop10', '--bind=/dev/loop11']unshare_net=TrueprintOutput=True) Using nspawn with args ['--capability=cap_ipc_lock', '--rlimit=RLIMIT_NOFILE=10240', '--capability=cap_ipc_lock', '--bind=/tmp/mock-resolv.ym2m_f69:/etc/resolv.conf', '--bind=/dev/btrfs-control', '--bind=/dev/loop-control', '--bind=/dev/loop0', '--bind=/dev/loop1', '--bind=/dev/loop2', '--bind=/dev/loop3', '--bind=/dev/loop4', '--bind=/dev/loop5', '--bind=/dev/loop6', '--bind=/dev/loop7', '--bind=/dev/loop8', '--bind=/dev/loop9', '--bind=/dev/loop10', '--bind=/dev/loop11'] Executing command: ['/usr/bin/systemd-nspawn', '-q', '-M', '31b0a562923644b9b346e4f8f0cfb1e2', '-D', '/var/lib/mock/fedora-rawhide-x86_64-1649687776.743359/root', '-a', '-u', 'mockbuild', '--capability=cap_ipc_lock', '--rlimit=RLIMIT_NOFILE=10240', '--capability=cap_ipc_lock', '--bind=/tmp/mock-resolv.ym2m_f69:/etc/resolv.conf', '--bind=/dev/btrfs-control', '--bind=/dev/loop-control', '--bind=/dev/loop0', '--bind=/dev/loop1', '--bind=/dev/loop2', '--bind=/dev/loop3', '--bind=/dev/loop4', '--bind=/dev/loop5', '--bind=/dev/loop6', '--bind=/dev/loop7', '--bind=/dev/loop8', '--bind=/dev/loop9', '--bind=/dev/loop10', '--bind=/dev/loop11', '--console=pipe', '--setenv=TERM=vt100', '--setenv=SHELL=/bin/bash', '--setenv=HOME=/builddir', '--setenv=HOSTNAME=mock', '--setenv=PATH=/usr/bin:/bin:/usr/sbin:/sbin', '--setenv=PROMPT_COMMAND=printf "\\033]0;\\007"', '--setenv=PS1= \\s-\\v\\$ ', '--setenv=LANG=C.UTF-8', '--resolv-conf=off', 'bash', '--login', '-c', '/usr/bin/rpmbuild -bs --target x86_64 --nodeps /builddir/build/SPECS/python-scrapy-dynamic-spiders.spec'] with env {'TERM': 'vt100', 'SHELL': '/bin/bash', 'HOME': '/builddir', 'HOSTNAME': 'mock', 'PATH': '/usr/bin:/bin:/usr/sbin:/sbin', 'PROMPT_COMMAND': 'printf "\\033]0;\\007"', 'PS1': ' \\s-\\v\\$ ', 'LANG': 'C.UTF-8', 'SYSTEMD_NSPAWN_TMPFS_TMP': '0', 'SYSTEMD_SECCOMP': '0'} and shell False Building target platforms: x86_64 Building for target x86_64 setting SOURCE_DATE_EPOCH=1649462400 Wrote: /builddir/build/SRPMS/python-scrapy-dynamic-spiders-1.0.0~a1-1.fc37.src.rpm Child return code was: 0 ENTER ['do_with_status'](['bash', '--login', '-c', '/usr/bin/rpmbuild -br --target x86_64 --nodeps /builddir/build/SPECS/python-scrapy-dynamic-spiders.spec'], chrootPath='/var/lib/mock/fedora-rawhide-x86_64-1649687776.743359/root'env={'TERM': 'vt100', 'SHELL': '/bin/bash', 'HOME': '/builddir', 'HOSTNAME': 'mock', 'PATH': '/usr/bin:/bin:/usr/sbin:/sbin', 'PROMPT_COMMAND': 'printf "\\033]0;\\007"', 'PS1': ' \\s-\\v\\$ ', 'LANG': 'C.UTF-8'}shell=Falselogger=timeout=0uid=1001gid=135user='mockbuild'nspawn_args=['--capability=cap_ipc_lock', '--rlimit=RLIMIT_NOFILE=10240', '--capability=cap_ipc_lock', '--bind=/tmp/mock-resolv.ym2m_f69:/etc/resolv.conf', '--bind=/dev/btrfs-control', '--bind=/dev/loop-control', '--bind=/dev/loop0', '--bind=/dev/loop1', '--bind=/dev/loop2', '--bind=/dev/loop3', '--bind=/dev/loop4', '--bind=/dev/loop5', '--bind=/dev/loop6', '--bind=/dev/loop7', '--bind=/dev/loop8', '--bind=/dev/loop9', '--bind=/dev/loop10', '--bind=/dev/loop11']unshare_net=TrueraiseExc=FalseprintOutput=True) Using nspawn with args ['--capability=cap_ipc_lock', '--rlimit=RLIMIT_NOFILE=10240', '--capability=cap_ipc_lock', '--bind=/tmp/mock-resolv.ym2m_f69:/etc/resolv.conf', '--bind=/dev/btrfs-control', '--bind=/dev/loop-control', '--bind=/dev/loop0', '--bind=/dev/loop1', '--bind=/dev/loop2', '--bind=/dev/loop3', '--bind=/dev/loop4', '--bind=/dev/loop5', '--bind=/dev/loop6', '--bind=/dev/loop7', '--bind=/dev/loop8', '--bind=/dev/loop9', '--bind=/dev/loop10', '--bind=/dev/loop11'] Executing command: ['/usr/bin/systemd-nspawn', '-q', '-M', 'e597941b2a5e49ecb5d0b016079b08c9', '-D', '/var/lib/mock/fedora-rawhide-x86_64-1649687776.743359/root', '-a', '-u', 'mockbuild', '--capability=cap_ipc_lock', '--rlimit=RLIMIT_NOFILE=10240', '--capability=cap_ipc_lock', '--bind=/tmp/mock-resolv.ym2m_f69:/etc/resolv.conf', '--bind=/dev/btrfs-control', '--bind=/dev/loop-control', '--bind=/dev/loop0', '--bind=/dev/loop1', '--bind=/dev/loop2', '--bind=/dev/loop3', '--bind=/dev/loop4', '--bind=/dev/loop5', '--bind=/dev/loop6', '--bind=/dev/loop7', '--bind=/dev/loop8', '--bind=/dev/loop9', '--bind=/dev/loop10', '--bind=/dev/loop11', '--console=pipe', '--setenv=TERM=vt100', '--setenv=SHELL=/bin/bash', '--setenv=HOME=/builddir', '--setenv=HOSTNAME=mock', '--setenv=PATH=/usr/bin:/bin:/usr/sbin:/sbin', '--setenv=PROMPT_COMMAND=printf "\\033]0;\\007"', '--setenv=PS1= \\s-\\v\\$ ', '--setenv=LANG=C.UTF-8', '--resolv-conf=off', 'bash', '--login', '-c', '/usr/bin/rpmbuild -br --target x86_64 --nodeps /builddir/build/SPECS/python-scrapy-dynamic-spiders.spec'] with env {'TERM': 'vt100', 'SHELL': '/bin/bash', 'HOME': '/builddir', 'HOSTNAME': 'mock', 'PATH': '/usr/bin:/bin:/usr/sbin:/sbin', 'PROMPT_COMMAND': 'printf "\\033]0;\\007"', 'PS1': ' \\s-\\v\\$ ', 'LANG': 'C.UTF-8', 'SYSTEMD_NSPAWN_TMPFS_TMP': '0', 'SYSTEMD_SECCOMP': '0'} and shell False Building target platforms: x86_64 Building for target x86_64 setting SOURCE_DATE_EPOCH=1649462400 Executing(%prep): /bin/sh -e /var/tmp/rpm-tmp.C4Wewq + umask 022 + cd /builddir/build/BUILD + cd /builddir/build/BUILD + rm -rf scrapy_dynamic_spiders-1.0.0a1 + /usr/bin/gzip -dc /builddir/build/SOURCES/scrapy_dynamic_spiders-1.0.0a1.tar.gz + /usr/bin/tar -xof - + STATUS=0 + '[' 0 -ne 0 ']' + cd scrapy_dynamic_spiders-1.0.0a1 + /usr/bin/chmod -Rf a+rX,u+w,g-w,o-w . + RPM_EC=0 ++ jobs -p + exit 0 Executing(%generate_buildrequires): /bin/sh -e /var/tmp/rpm-tmp.MegP8j + umask 022 + cd /builddir/build/BUILD + cd scrapy_dynamic_spiders-1.0.0a1 + echo pyproject-rpm-macros + echo python3-devel + echo 'python3dist(pip) >= 19' + echo 'python3dist(packaging)' + '[' -f pyproject.toml ']' + '[' -f setup.py ']' + echo 'python3dist(setuptools) >= 40.8' + echo 'python3dist(wheel)' + rm -rfv '*.dist-info/' + '[' -f /usr/bin/python3 ']' + RPM_TOXENV=py310 + HOSTNAME=rpmbuild + /usr/bin/python3 -s /usr/lib/rpm/redhat/pyproject_buildrequires.py --generate-extras --python3_pkgversion 3 Handling setuptools >= 40.8 from default build backend Requirement not satisfied: setuptools >= 40.8 Handling wheel from default build backend Requirement not satisfied: wheel Exiting dependency generation pass: build backend + RPM_EC=0 ++ jobs -p + exit 0 Wrote: /builddir/build/SRPMS/python-scrapy-dynamic-spiders-1.0.0~a1-1.fc37.buildreqs.nosrc.rpm Child return code was: 11 Dynamic buildrequires detected Going to install missing buildrequires. See root.log for details. ENTER ['do_with_status'](['bash', '--login', '-c', '/usr/bin/rpmbuild -br --target x86_64 --nodeps /builddir/build/SPECS/python-scrapy-dynamic-spiders.spec'], chrootPath='/var/lib/mock/fedora-rawhide-x86_64-1649687776.743359/root'env={'TERM': 'vt100', 'SHELL': '/bin/bash', 'HOME': '/builddir', 'HOSTNAME': 'mock', 'PATH': '/usr/bin:/bin:/usr/sbin:/sbin', 'PROMPT_COMMAND': 'printf "\\033]0;\\007"', 'PS1': ' \\s-\\v\\$ ', 'LANG': 'C.UTF-8'}shell=Falselogger=timeout=0uid=1001gid=135user='mockbuild'nspawn_args=['--capability=cap_ipc_lock', '--rlimit=RLIMIT_NOFILE=10240', '--capability=cap_ipc_lock', '--bind=/tmp/mock-resolv.ym2m_f69:/etc/resolv.conf', '--bind=/dev/btrfs-control', '--bind=/dev/loop-control', '--bind=/dev/loop0', '--bind=/dev/loop1', '--bind=/dev/loop2', '--bind=/dev/loop3', '--bind=/dev/loop4', '--bind=/dev/loop5', '--bind=/dev/loop6', '--bind=/dev/loop7', '--bind=/dev/loop8', '--bind=/dev/loop9', '--bind=/dev/loop10', '--bind=/dev/loop11']unshare_net=TrueraiseExc=FalseprintOutput=True) Using nspawn with args ['--capability=cap_ipc_lock', '--rlimit=RLIMIT_NOFILE=10240', '--capability=cap_ipc_lock', '--bind=/tmp/mock-resolv.ym2m_f69:/etc/resolv.conf', '--bind=/dev/btrfs-control', '--bind=/dev/loop-control', '--bind=/dev/loop0', '--bind=/dev/loop1', '--bind=/dev/loop2', '--bind=/dev/loop3', '--bind=/dev/loop4', '--bind=/dev/loop5', '--bind=/dev/loop6', '--bind=/dev/loop7', '--bind=/dev/loop8', '--bind=/dev/loop9', '--bind=/dev/loop10', '--bind=/dev/loop11'] Executing command: ['/usr/bin/systemd-nspawn', '-q', '-M', 'e87b4628a4fc49d8be4b76b19b88c25c', '-D', '/var/lib/mock/fedora-rawhide-x86_64-1649687776.743359/root', '-a', '-u', 'mockbuild', '--capability=cap_ipc_lock', '--rlimit=RLIMIT_NOFILE=10240', '--capability=cap_ipc_lock', '--bind=/tmp/mock-resolv.ym2m_f69:/etc/resolv.conf', '--bind=/dev/btrfs-control', '--bind=/dev/loop-control', '--bind=/dev/loop0', '--bind=/dev/loop1', '--bind=/dev/loop2', '--bind=/dev/loop3', '--bind=/dev/loop4', '--bind=/dev/loop5', '--bind=/dev/loop6', '--bind=/dev/loop7', '--bind=/dev/loop8', '--bind=/dev/loop9', '--bind=/dev/loop10', '--bind=/dev/loop11', '--console=pipe', '--setenv=TERM=vt100', '--setenv=SHELL=/bin/bash', '--setenv=HOME=/builddir', '--setenv=HOSTNAME=mock', '--setenv=PATH=/usr/bin:/bin:/usr/sbin:/sbin', '--setenv=PROMPT_COMMAND=printf "\\033]0;\\007"', '--setenv=PS1= \\s-\\v\\$ ', '--setenv=LANG=C.UTF-8', '--resolv-conf=off', 'bash', '--login', '-c', '/usr/bin/rpmbuild -br --target x86_64 --nodeps /builddir/build/SPECS/python-scrapy-dynamic-spiders.spec'] with env {'TERM': 'vt100', 'SHELL': '/bin/bash', 'HOME': '/builddir', 'HOSTNAME': 'mock', 'PATH': '/usr/bin:/bin:/usr/sbin:/sbin', 'PROMPT_COMMAND': 'printf "\\033]0;\\007"', 'PS1': ' \\s-\\v\\$ ', 'LANG': 'C.UTF-8', 'SYSTEMD_NSPAWN_TMPFS_TMP': '0', 'SYSTEMD_SECCOMP': '0'} and shell False Building target platforms: x86_64 Building for target x86_64 setting SOURCE_DATE_EPOCH=1649462400 Executing(%prep): /bin/sh -e /var/tmp/rpm-tmp.0Tg4so + umask 022 + cd /builddir/build/BUILD + cd /builddir/build/BUILD + rm -rf scrapy_dynamic_spiders-1.0.0a1 + /usr/bin/gzip -dc /builddir/build/SOURCES/scrapy_dynamic_spiders-1.0.0a1.tar.gz + /usr/bin/tar -xof - + STATUS=0 + '[' 0 -ne 0 ']' + cd scrapy_dynamic_spiders-1.0.0a1 + /usr/bin/chmod -Rf a+rX,u+w,g-w,o-w . + RPM_EC=0 ++ jobs -p + exit 0 Executing(%generate_buildrequires): /bin/sh -e /var/tmp/rpm-tmp.OgERmP + umask 022 + cd /builddir/build/BUILD + cd scrapy_dynamic_spiders-1.0.0a1 + echo pyproject-rpm-macros + echo python3-devel + echo 'python3dist(pip) >= 19' + echo 'python3dist(packaging)' + '[' -f pyproject.toml ']' + '[' -f setup.py ']' + echo 'python3dist(setuptools) >= 40.8' + echo 'python3dist(wheel)' + rm -rfv '*.dist-info/' + '[' -f /usr/bin/python3 ']' + RPM_TOXENV=py310 + HOSTNAME=rpmbuild + /usr/bin/python3 -s /usr/lib/rpm/redhat/pyproject_buildrequires.py --generate-extras --python3_pkgversion 3 Handling setuptools >= 40.8 from default build backend Requirement satisfied: setuptools >= 40.8 (installed: setuptools 60.9.3) Handling wheel from default build backend Requirement satisfied: wheel (installed: wheel 0.37.1) /usr/lib/python3.10/site-packages/setuptools/dist.py:505: UserWarning: Normalizing '1.0.0.a1' to '1.0.0a1' warnings.warn(tmpl.format(**locals())) HOOK STDOUT: running egg_info HOOK STDOUT: writing scrapy_dynamic_spiders.egg-info/PKG-INFO HOOK STDOUT: writing dependency_links to scrapy_dynamic_spiders.egg-info/dependency_links.txt HOOK STDOUT: writing requirements to scrapy_dynamic_spiders.egg-info/requires.txt HOOK STDOUT: writing top-level names to scrapy_dynamic_spiders.egg-info/top_level.txt HOOK STDOUT: reading manifest file 'scrapy_dynamic_spiders.egg-info/SOURCES.txt' HOOK STDOUT: writing manifest file 'scrapy_dynamic_spiders.egg-info/SOURCES.txt' Handling wheel from get_requires_for_build_wheel Requirement satisfied: wheel (installed: wheel 0.37.1) HOOK STDOUT: running dist_info HOOK STDOUT: writing scrapy_dynamic_spiders.egg-info/PKG-INFO HOOK STDOUT: writing dependency_links to scrapy_dynamic_spiders.egg-info/dependency_links.txt HOOK STDOUT: writing requirements to scrapy_dynamic_spiders.egg-info/requires.txt HOOK STDOUT: writing top-level names to scrapy_dynamic_spiders.egg-info/top_level.txt HOOK STDOUT: reading manifest file 'scrapy_dynamic_spiders.egg-info/SOURCES.txt' HOOK STDOUT: writing manifest file 'scrapy_dynamic_spiders.egg-info/SOURCES.txt' HOOK STDOUT: creating '/builddir/build/BUILD/scrapy_dynamic_spiders-1.0.0a1/scrapy_dynamic_spiders.dist-info' Handling scrapy from wheel metadata: Requires-Dist Requirement not satisfied: scrapy Handling crochet from wheel metadata: Requires-Dist Requirement not satisfied: crochet + RPM_EC=0 ++ jobs -p + exit 0 Wrote: /builddir/build/SRPMS/python-scrapy-dynamic-spiders-1.0.0~a1-1.fc37.buildreqs.nosrc.rpm Child return code was: 11 Dynamic buildrequires detected Going to install missing buildrequires. See root.log for details. ENTER ['do_with_status'](['bash', '--login', '-c', '/usr/bin/rpmbuild -br --target x86_64 --nodeps /builddir/build/SPECS/python-scrapy-dynamic-spiders.spec'], chrootPath='/var/lib/mock/fedora-rawhide-x86_64-1649687776.743359/root'env={'TERM': 'vt100', 'SHELL': '/bin/bash', 'HOME': '/builddir', 'HOSTNAME': 'mock', 'PATH': '/usr/bin:/bin:/usr/sbin:/sbin', 'PROMPT_COMMAND': 'printf "\\033]0;\\007"', 'PS1': ' \\s-\\v\\$ ', 'LANG': 'C.UTF-8'}shell=Falselogger=timeout=0uid=1001gid=135user='mockbuild'nspawn_args=['--capability=cap_ipc_lock', '--rlimit=RLIMIT_NOFILE=10240', '--capability=cap_ipc_lock', '--bind=/tmp/mock-resolv.ym2m_f69:/etc/resolv.conf', '--bind=/dev/btrfs-control', '--bind=/dev/loop-control', '--bind=/dev/loop0', '--bind=/dev/loop1', '--bind=/dev/loop2', '--bind=/dev/loop3', '--bind=/dev/loop4', '--bind=/dev/loop5', '--bind=/dev/loop6', '--bind=/dev/loop7', '--bind=/dev/loop8', '--bind=/dev/loop9', '--bind=/dev/loop10', '--bind=/dev/loop11']unshare_net=TrueraiseExc=FalseprintOutput=True) Using nspawn with args ['--capability=cap_ipc_lock', '--rlimit=RLIMIT_NOFILE=10240', '--capability=cap_ipc_lock', '--bind=/tmp/mock-resolv.ym2m_f69:/etc/resolv.conf', '--bind=/dev/btrfs-control', '--bind=/dev/loop-control', '--bind=/dev/loop0', '--bind=/dev/loop1', '--bind=/dev/loop2', '--bind=/dev/loop3', '--bind=/dev/loop4', '--bind=/dev/loop5', '--bind=/dev/loop6', '--bind=/dev/loop7', '--bind=/dev/loop8', '--bind=/dev/loop9', '--bind=/dev/loop10', '--bind=/dev/loop11'] Executing command: ['/usr/bin/systemd-nspawn', '-q', '-M', 'ff0e61b6b03d45008a0fab3d374ccd12', '-D', '/var/lib/mock/fedora-rawhide-x86_64-1649687776.743359/root', '-a', '-u', 'mockbuild', '--capability=cap_ipc_lock', '--rlimit=RLIMIT_NOFILE=10240', '--capability=cap_ipc_lock', '--bind=/tmp/mock-resolv.ym2m_f69:/etc/resolv.conf', '--bind=/dev/btrfs-control', '--bind=/dev/loop-control', '--bind=/dev/loop0', '--bind=/dev/loop1', '--bind=/dev/loop2', '--bind=/dev/loop3', '--bind=/dev/loop4', '--bind=/dev/loop5', '--bind=/dev/loop6', '--bind=/dev/loop7', '--bind=/dev/loop8', '--bind=/dev/loop9', '--bind=/dev/loop10', '--bind=/dev/loop11', '--console=pipe', '--setenv=TERM=vt100', '--setenv=SHELL=/bin/bash', '--setenv=HOME=/builddir', '--setenv=HOSTNAME=mock', '--setenv=PATH=/usr/bin:/bin:/usr/sbin:/sbin', '--setenv=PROMPT_COMMAND=printf "\\033]0;\\007"', '--setenv=PS1= \\s-\\v\\$ ', '--setenv=LANG=C.UTF-8', '--resolv-conf=off', 'bash', '--login', '-c', '/usr/bin/rpmbuild -br --target x86_64 --nodeps /builddir/build/SPECS/python-scrapy-dynamic-spiders.spec'] with env {'TERM': 'vt100', 'SHELL': '/bin/bash', 'HOME': '/builddir', 'HOSTNAME': 'mock', 'PATH': '/usr/bin:/bin:/usr/sbin:/sbin', 'PROMPT_COMMAND': 'printf "\\033]0;\\007"', 'PS1': ' \\s-\\v\\$ ', 'LANG': 'C.UTF-8', 'SYSTEMD_NSPAWN_TMPFS_TMP': '0', 'SYSTEMD_SECCOMP': '0'} and shell False Building target platforms: x86_64 Building for target x86_64 setting SOURCE_DATE_EPOCH=1649462400 Executing(%prep): /bin/sh -e /var/tmp/rpm-tmp.vK5G4S + umask 022 + cd /builddir/build/BUILD + cd /builddir/build/BUILD + rm -rf scrapy_dynamic_spiders-1.0.0a1 + /usr/bin/gzip -dc /builddir/build/SOURCES/scrapy_dynamic_spiders-1.0.0a1.tar.gz + /usr/bin/tar -xof - + STATUS=0 + '[' 0 -ne 0 ']' + cd scrapy_dynamic_spiders-1.0.0a1 + /usr/bin/chmod -Rf a+rX,u+w,g-w,o-w . + RPM_EC=0 ++ jobs -p + exit 0 Executing(%generate_buildrequires): /bin/sh -e /var/tmp/rpm-tmp.SlEhTk + umask 022 + cd /builddir/build/BUILD + cd scrapy_dynamic_spiders-1.0.0a1 + echo pyproject-rpm-macros + echo python3-devel + echo 'python3dist(pip) >= 19' + echo 'python3dist(packaging)' + '[' -f pyproject.toml ']' + '[' -f setup.py ']' + echo 'python3dist(setuptools) >= 40.8' + echo 'python3dist(wheel)' + rm -rfv '*.dist-info/' + '[' -f /usr/bin/python3 ']' + RPM_TOXENV=py310 + HOSTNAME=rpmbuild + /usr/bin/python3 -s /usr/lib/rpm/redhat/pyproject_buildrequires.py --generate-extras --python3_pkgversion 3 Handling setuptools >= 40.8 from default build backend Requirement satisfied: setuptools >= 40.8 (installed: setuptools 60.9.3) Handling wheel from default build backend Requirement satisfied: wheel (installed: wheel 0.37.1) /usr/lib/python3.10/site-packages/setuptools/dist.py:505: UserWarning: Normalizing '1.0.0.a1' to '1.0.0a1' warnings.warn(tmpl.format(**locals())) HOOK STDOUT: running egg_info HOOK STDOUT: writing scrapy_dynamic_spiders.egg-info/PKG-INFO HOOK STDOUT: writing dependency_links to scrapy_dynamic_spiders.egg-info/dependency_links.txt HOOK STDOUT: writing requirements to scrapy_dynamic_spiders.egg-info/requires.txt HOOK STDOUT: writing top-level names to scrapy_dynamic_spiders.egg-info/top_level.txt HOOK STDOUT: reading manifest file 'scrapy_dynamic_spiders.egg-info/SOURCES.txt' HOOK STDOUT: writing manifest file 'scrapy_dynamic_spiders.egg-info/SOURCES.txt' Handling wheel from get_requires_for_build_wheel Requirement satisfied: wheel (installed: wheel 0.37.1) HOOK STDOUT: running dist_info HOOK STDOUT: writing scrapy_dynamic_spiders.egg-info/PKG-INFO HOOK STDOUT: writing dependency_links to scrapy_dynamic_spiders.egg-info/dependency_links.txt HOOK STDOUT: writing requirements to scrapy_dynamic_spiders.egg-info/requires.txt HOOK STDOUT: writing top-level names to scrapy_dynamic_spiders.egg-info/top_level.txt HOOK STDOUT: reading manifest file 'scrapy_dynamic_spiders.egg-info/SOURCES.txt' HOOK STDOUT: writing manifest file 'scrapy_dynamic_spiders.egg-info/SOURCES.txt' HOOK STDOUT: creating '/builddir/build/BUILD/scrapy_dynamic_spiders-1.0.0a1/scrapy_dynamic_spiders.dist-info' Handling scrapy from wheel metadata: Requires-Dist Requirement satisfied: scrapy (installed: scrapy 2.6.1) Handling crochet from wheel metadata: Requires-Dist Requirement satisfied: crochet (installed: crochet 2.0.0) + RPM_EC=0 ++ jobs -p + exit 0 Wrote: /builddir/build/SRPMS/python-scrapy-dynamic-spiders-1.0.0~a1-1.fc37.buildreqs.nosrc.rpm Child return code was: 11 Dynamic buildrequires detected Going to install missing buildrequires. See root.log for details. ENTER ['do_with_status'](['bash', '--login', '-c', '/usr/bin/rpmbuild -ba --noprep --target x86_64 --nodeps /builddir/build/SPECS/python-scrapy-dynamic-spiders.spec'], chrootPath='/var/lib/mock/fedora-rawhide-x86_64-1649687776.743359/root'env={'TERM': 'vt100', 'SHELL': '/bin/bash', 'HOME': '/builddir', 'HOSTNAME': 'mock', 'PATH': '/usr/bin:/bin:/usr/sbin:/sbin', 'PROMPT_COMMAND': 'printf "\\033]0;\\007"', 'PS1': ' \\s-\\v\\$ ', 'LANG': 'C.UTF-8'}shell=Falselogger=timeout=0uid=1001gid=135user='mockbuild'nspawn_args=['--capability=cap_ipc_lock', '--rlimit=RLIMIT_NOFILE=10240', '--capability=cap_ipc_lock', '--bind=/tmp/mock-resolv.ym2m_f69:/etc/resolv.conf', '--bind=/dev/btrfs-control', '--bind=/dev/loop-control', '--bind=/dev/loop0', '--bind=/dev/loop1', '--bind=/dev/loop2', '--bind=/dev/loop3', '--bind=/dev/loop4', '--bind=/dev/loop5', '--bind=/dev/loop6', '--bind=/dev/loop7', '--bind=/dev/loop8', '--bind=/dev/loop9', '--bind=/dev/loop10', '--bind=/dev/loop11']unshare_net=TrueprintOutput=True) Using nspawn with args ['--capability=cap_ipc_lock', '--rlimit=RLIMIT_NOFILE=10240', '--capability=cap_ipc_lock', '--bind=/tmp/mock-resolv.ym2m_f69:/etc/resolv.conf', '--bind=/dev/btrfs-control', '--bind=/dev/loop-control', '--bind=/dev/loop0', '--bind=/dev/loop1', '--bind=/dev/loop2', '--bind=/dev/loop3', '--bind=/dev/loop4', '--bind=/dev/loop5', '--bind=/dev/loop6', '--bind=/dev/loop7', '--bind=/dev/loop8', '--bind=/dev/loop9', '--bind=/dev/loop10', '--bind=/dev/loop11'] Executing command: ['/usr/bin/systemd-nspawn', '-q', '-M', '3706dca694d341698e81553630602b66', '-D', '/var/lib/mock/fedora-rawhide-x86_64-1649687776.743359/root', '-a', '-u', 'mockbuild', '--capability=cap_ipc_lock', '--rlimit=RLIMIT_NOFILE=10240', '--capability=cap_ipc_lock', '--bind=/tmp/mock-resolv.ym2m_f69:/etc/resolv.conf', '--bind=/dev/btrfs-control', '--bind=/dev/loop-control', '--bind=/dev/loop0', '--bind=/dev/loop1', '--bind=/dev/loop2', '--bind=/dev/loop3', '--bind=/dev/loop4', '--bind=/dev/loop5', '--bind=/dev/loop6', '--bind=/dev/loop7', '--bind=/dev/loop8', '--bind=/dev/loop9', '--bind=/dev/loop10', '--bind=/dev/loop11', '--console=pipe', '--setenv=TERM=vt100', '--setenv=SHELL=/bin/bash', '--setenv=HOME=/builddir', '--setenv=HOSTNAME=mock', '--setenv=PATH=/usr/bin:/bin:/usr/sbin:/sbin', '--setenv=PROMPT_COMMAND=printf "\\033]0;\\007"', '--setenv=PS1= \\s-\\v\\$ ', '--setenv=LANG=C.UTF-8', '--resolv-conf=off', 'bash', '--login', '-c', '/usr/bin/rpmbuild -ba --noprep --target x86_64 --nodeps /builddir/build/SPECS/python-scrapy-dynamic-spiders.spec'] with env {'TERM': 'vt100', 'SHELL': '/bin/bash', 'HOME': '/builddir', 'HOSTNAME': 'mock', 'PATH': '/usr/bin:/bin:/usr/sbin:/sbin', 'PROMPT_COMMAND': 'printf "\\033]0;\\007"', 'PS1': ' \\s-\\v\\$ ', 'LANG': 'C.UTF-8', 'SYSTEMD_NSPAWN_TMPFS_TMP': '0', 'SYSTEMD_SECCOMP': '0'} and shell False Building target platforms: x86_64 Building for target x86_64 setting SOURCE_DATE_EPOCH=1649462400 Executing(%generate_buildrequires): /bin/sh -e /var/tmp/rpm-tmp.lcCi48 + umask 022 + cd /builddir/build/BUILD + cd scrapy_dynamic_spiders-1.0.0a1 + echo pyproject-rpm-macros + echo python3-devel + echo 'python3dist(pip) >= 19' + echo 'python3dist(packaging)' + '[' -f pyproject.toml ']' + '[' -f setup.py ']' + echo 'python3dist(setuptools) >= 40.8' + echo 'python3dist(wheel)' + rm -rfv scrapy_dynamic_spiders.dist-info/ removed 'scrapy_dynamic_spiders.dist-info/METADATA' removed 'scrapy_dynamic_spiders.dist-info/top_level.txt' removed directory 'scrapy_dynamic_spiders.dist-info/' + '[' -f /usr/bin/python3 ']' + RPM_TOXENV=py310 + HOSTNAME=rpmbuild + /usr/bin/python3 -s /usr/lib/rpm/redhat/pyproject_buildrequires.py --generate-extras --python3_pkgversion 3 Handling setuptools >= 40.8 from default build backend Requirement satisfied: setuptools >= 40.8 (installed: setuptools 60.9.3) Handling wheel from default build backend Requirement satisfied: wheel (installed: wheel 0.37.1) /usr/lib/python3.10/site-packages/setuptools/dist.py:505: UserWarning: Normalizing '1.0.0.a1' to '1.0.0a1' warnings.warn(tmpl.format(**locals())) HOOK STDOUT: running egg_info HOOK STDOUT: creating scrapy_dynamic_spiders.egg-info HOOK STDOUT: writing scrapy_dynamic_spiders.egg-info/PKG-INFO HOOK STDOUT: writing dependency_links to scrapy_dynamic_spiders.egg-info/dependency_links.txt HOOK STDOUT: writing requirements to scrapy_dynamic_spiders.egg-info/requires.txt HOOK STDOUT: writing top-level names to scrapy_dynamic_spiders.egg-info/top_level.txt HOOK STDOUT: writing manifest file 'scrapy_dynamic_spiders.egg-info/SOURCES.txt' HOOK STDOUT: reading manifest file 'scrapy_dynamic_spiders.egg-info/SOURCES.txt' HOOK STDOUT: writing manifest file 'scrapy_dynamic_spiders.egg-info/SOURCES.txt' Handling wheel from get_requires_for_build_wheel Requirement satisfied: wheel (installed: wheel 0.37.1) HOOK STDOUT: running dist_info HOOK STDOUT: writing scrapy_dynamic_spiders.egg-info/PKG-INFO HOOK STDOUT: writing dependency_links to scrapy_dynamic_spiders.egg-info/dependency_links.txt HOOK STDOUT: writing requirements to scrapy_dynamic_spiders.egg-info/requires.txt HOOK STDOUT: writing top-level names to scrapy_dynamic_spiders.egg-info/top_level.txt HOOK STDOUT: reading manifest file 'scrapy_dynamic_spiders.egg-info/SOURCES.txt' HOOK STDOUT: writing manifest file 'scrapy_dynamic_spiders.egg-info/SOURCES.txt' HOOK STDOUT: creating '/builddir/build/BUILD/scrapy_dynamic_spiders-1.0.0a1/scrapy_dynamic_spiders.dist-info' Handling scrapy from wheel metadata: Requires-Dist Requirement satisfied: scrapy (installed: scrapy 2.6.1) Handling crochet from wheel metadata: Requires-Dist Requirement satisfied: crochet (installed: crochet 2.0.0) + RPM_EC=0 ++ jobs -p + exit 0 Executing(%build): /bin/sh -e /var/tmp/rpm-tmp.ESpvUu + umask 022 + cd /builddir/build/BUILD + CFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection' + export CFLAGS + CXXFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection' + export CXXFLAGS + FFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -I/usr/lib64/gfortran/modules' + export FFLAGS + FCFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -I/usr/lib64/gfortran/modules' + export FCFLAGS + LDFLAGS='-Wl,-z,relro -Wl,--as-needed -Wl,-z,now -specs=/usr/lib/rpm/redhat/redhat-hardened-ld -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -Wl,--build-id=sha1 -Wl,-dT,/builddir/build/BUILD/scrapy_dynamic_spiders-1.0.0a1/.package_note-python-scrapy-dynamic-spiders-1.0.0~a1-1.fc37.x86_64.ld' + export LDFLAGS + LT_SYS_LIBRARY_PATH=/usr/lib64: + export LT_SYS_LIBRARY_PATH + CC=gcc + export CC + CXX=g++ + export CXX + '[' -f /usr/lib/rpm/generate-rpm-note.sh ']' + /usr/lib/rpm/generate-rpm-note.sh python-scrapy-dynamic-spiders 1.0.0~a1-1.fc37 x86_64 + cd scrapy_dynamic_spiders-1.0.0a1 + mkdir -p /builddir/build/BUILD/scrapy_dynamic_spiders-1.0.0a1/.pyproject-builddir + CFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection' + LDFLAGS='-Wl,-z,relro -Wl,--as-needed -Wl,-z,now -specs=/usr/lib/rpm/redhat/redhat-hardened-ld -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -Wl,--build-id=sha1 -Wl,-dT,/builddir/build/BUILD/scrapy_dynamic_spiders-1.0.0a1/.package_note-python-scrapy-dynamic-spiders-1.0.0~a1-1.fc37.x86_64.ld' + TMPDIR=/builddir/build/BUILD/scrapy_dynamic_spiders-1.0.0a1/.pyproject-builddir + /usr/bin/python3 -m pip wheel --wheel-dir /builddir/build/BUILD/scrapy_dynamic_spiders-1.0.0a1/pyproject-wheeldir --no-deps --use-pep517 --no-build-isolation --disable-pip-version-check --no-clean --progress-bar off --verbose . Processing /builddir/build/BUILD/scrapy_dynamic_spiders-1.0.0a1 Preparing metadata (pyproject.toml): started Running command Preparing metadata (pyproject.toml) /usr/lib/python3.10/site-packages/setuptools/dist.py:505: UserWarning: Normalizing '1.0.0.a1' to '1.0.0a1' warnings.warn(tmpl.format(**locals())) running dist_info creating /builddir/build/BUILD/scrapy_dynamic_spiders-1.0.0a1/.pyproject-builddir/pip-modern-metadata-vzafsgj_/scrapy_dynamic_spiders.egg-info writing /builddir/build/BUILD/scrapy_dynamic_spiders-1.0.0a1/.pyproject-builddir/pip-modern-metadata-vzafsgj_/scrapy_dynamic_spiders.egg-info/PKG-INFO writing dependency_links to /builddir/build/BUILD/scrapy_dynamic_spiders-1.0.0a1/.pyproject-builddir/pip-modern-metadata-vzafsgj_/scrapy_dynamic_spiders.egg-info/dependency_links.txt writing requirements to /builddir/build/BUILD/scrapy_dynamic_spiders-1.0.0a1/.pyproject-builddir/pip-modern-metadata-vzafsgj_/scrapy_dynamic_spiders.egg-info/requires.txt writing top-level names to /builddir/build/BUILD/scrapy_dynamic_spiders-1.0.0a1/.pyproject-builddir/pip-modern-metadata-vzafsgj_/scrapy_dynamic_spiders.egg-info/top_level.txt writing manifest file '/builddir/build/BUILD/scrapy_dynamic_spiders-1.0.0a1/.pyproject-builddir/pip-modern-metadata-vzafsgj_/scrapy_dynamic_spiders.egg-info/SOURCES.txt' reading manifest file '/builddir/build/BUILD/scrapy_dynamic_spiders-1.0.0a1/.pyproject-builddir/pip-modern-metadata-vzafsgj_/scrapy_dynamic_spiders.egg-info/SOURCES.txt' writing manifest file '/builddir/build/BUILD/scrapy_dynamic_spiders-1.0.0a1/.pyproject-builddir/pip-modern-metadata-vzafsgj_/scrapy_dynamic_spiders.egg-info/SOURCES.txt' creating '/builddir/build/BUILD/scrapy_dynamic_spiders-1.0.0a1/.pyproject-builddir/pip-modern-metadata-vzafsgj_/scrapy_dynamic_spiders.dist-info' Preparing metadata (pyproject.toml): finished with status 'done' Building wheels for collected packages: scrapy-dynamic-spiders Building wheel for scrapy-dynamic-spiders (pyproject.toml): started Running command Building wheel for scrapy-dynamic-spiders (pyproject.toml) /usr/lib/python3.10/site-packages/setuptools/dist.py:505: UserWarning: Normalizing '1.0.0.a1' to '1.0.0a1' warnings.warn(tmpl.format(**locals())) running bdist_wheel running build running build_py creating build creating build/lib creating build/lib/tests copying tests/test_crawl_factory.py -> build/lib/tests copying tests/test_base_wrangler.py -> build/lib/tests copying tests/test_base_factory.py -> build/lib/tests copying tests/__init__.py -> build/lib/tests creating build/lib/scrapy_dynamic_spiders copying scrapy_dynamic_spiders/__init__.py -> build/lib/scrapy_dynamic_spiders creating build/lib/tests/testing_spiders copying tests/testing_spiders/spiders.py -> build/lib/tests/testing_spiders copying tests/testing_spiders/__init__.py -> build/lib/tests/testing_spiders creating build/lib/scrapy_dynamic_spiders/wranglers copying scrapy_dynamic_spiders/wranglers/base_wrangler.py -> build/lib/scrapy_dynamic_spiders/wranglers copying scrapy_dynamic_spiders/wranglers/__init__.py -> build/lib/scrapy_dynamic_spiders/wranglers creating build/lib/scrapy_dynamic_spiders/utils copying scrapy_dynamic_spiders/utils/factory_utils.py -> build/lib/scrapy_dynamic_spiders/utils copying scrapy_dynamic_spiders/utils/__init__.py -> build/lib/scrapy_dynamic_spiders/utils creating build/lib/scrapy_dynamic_spiders/factories copying scrapy_dynamic_spiders/factories/crawl_spider_factory.py -> build/lib/scrapy_dynamic_spiders/factories copying scrapy_dynamic_spiders/factories/base_factory.py -> build/lib/scrapy_dynamic_spiders/factories copying scrapy_dynamic_spiders/factories/__init__.py -> build/lib/scrapy_dynamic_spiders/factories installing to build/bdist.linux-x86_64/wheel running install running install_lib creating build/bdist.linux-x86_64 creating build/bdist.linux-x86_64/wheel creating build/bdist.linux-x86_64/wheel/scrapy_dynamic_spiders creating build/bdist.linux-x86_64/wheel/scrapy_dynamic_spiders/factories copying build/lib/scrapy_dynamic_spiders/factories/__init__.py -> build/bdist.linux-x86_64/wheel/scrapy_dynamic_spiders/factories copying build/lib/scrapy_dynamic_spiders/factories/base_factory.py -> build/bdist.linux-x86_64/wheel/scrapy_dynamic_spiders/factories copying build/lib/scrapy_dynamic_spiders/factories/crawl_spider_factory.py -> build/bdist.linux-x86_64/wheel/scrapy_dynamic_spiders/factories creating build/bdist.linux-x86_64/wheel/scrapy_dynamic_spiders/utils copying build/lib/scrapy_dynamic_spiders/utils/__init__.py -> build/bdist.linux-x86_64/wheel/scrapy_dynamic_spiders/utils copying build/lib/scrapy_dynamic_spiders/utils/factory_utils.py -> build/bdist.linux-x86_64/wheel/scrapy_dynamic_spiders/utils creating build/bdist.linux-x86_64/wheel/scrapy_dynamic_spiders/wranglers copying build/lib/scrapy_dynamic_spiders/wranglers/__init__.py -> build/bdist.linux-x86_64/wheel/scrapy_dynamic_spiders/wranglers copying build/lib/scrapy_dynamic_spiders/wranglers/base_wrangler.py -> build/bdist.linux-x86_64/wheel/scrapy_dynamic_spiders/wranglers copying build/lib/scrapy_dynamic_spiders/__init__.py -> build/bdist.linux-x86_64/wheel/scrapy_dynamic_spiders creating build/bdist.linux-x86_64/wheel/tests creating build/bdist.linux-x86_64/wheel/tests/testing_spiders copying build/lib/tests/testing_spiders/__init__.py -> build/bdist.linux-x86_64/wheel/tests/testing_spiders copying build/lib/tests/testing_spiders/spiders.py -> build/bdist.linux-x86_64/wheel/tests/testing_spiders copying build/lib/tests/__init__.py -> build/bdist.linux-x86_64/wheel/tests copying build/lib/tests/test_base_factory.py -> build/bdist.linux-x86_64/wheel/tests copying build/lib/tests/test_base_wrangler.py -> build/bdist.linux-x86_64/wheel/tests copying build/lib/tests/test_crawl_factory.py -> build/bdist.linux-x86_64/wheel/tests running install_egg_info running egg_info creating scrapy_dynamic_spiders.egg-info writing scrapy_dynamic_spiders.egg-info/PKG-INFO writing dependency_links to scrapy_dynamic_spiders.egg-info/dependency_links.txt writing requirements to scrapy_dynamic_spiders.egg-info/requires.txt writing top-level names to scrapy_dynamic_spiders.egg-info/top_level.txt writing manifest file 'scrapy_dynamic_spiders.egg-info/SOURCES.txt' reading manifest file 'scrapy_dynamic_spiders.egg-info/SOURCES.txt' writing manifest file 'scrapy_dynamic_spiders.egg-info/SOURCES.txt' Copying scrapy_dynamic_spiders.egg-info to build/bdist.linux-x86_64/wheel/scrapy_dynamic_spiders-1.0.0a1-py3.10.egg-info running install_scripts creating build/bdist.linux-x86_64/wheel/scrapy_dynamic_spiders-1.0.0a1.dist-info/WHEEL creating '/builddir/build/BUILD/scrapy_dynamic_spiders-1.0.0a1/.pyproject-builddir/pip-wheel-56s1ssyo/tmpifxtsibt/scrapy_dynamic_spiders-1.0.0a1-py3-none-any.whl' and adding 'build/bdist.linux-x86_64/wheel' to it adding 'scrapy_dynamic_spiders/__init__.py' adding 'scrapy_dynamic_spiders/factories/__init__.py' adding 'scrapy_dynamic_spiders/factories/base_factory.py' adding 'scrapy_dynamic_spiders/factories/crawl_spider_factory.py' adding 'scrapy_dynamic_spiders/utils/__init__.py' adding 'scrapy_dynamic_spiders/utils/factory_utils.py' adding 'scrapy_dynamic_spiders/wranglers/__init__.py' adding 'scrapy_dynamic_spiders/wranglers/base_wrangler.py' adding 'tests/__init__.py' adding 'tests/test_base_factory.py' adding 'tests/test_base_wrangler.py' adding 'tests/test_crawl_factory.py' adding 'tests/testing_spiders/__init__.py' adding 'tests/testing_spiders/spiders.py' adding 'scrapy_dynamic_spiders-1.0.0a1.dist-info/METADATA' adding 'scrapy_dynamic_spiders-1.0.0a1.dist-info/WHEEL' adding 'scrapy_dynamic_spiders-1.0.0a1.dist-info/top_level.txt' adding 'scrapy_dynamic_spiders-1.0.0a1.dist-info/RECORD' removing build/bdist.linux-x86_64/wheel Building wheel for scrapy-dynamic-spiders (pyproject.toml): finished with status 'done' Created wheel for scrapy-dynamic-spiders: filename=scrapy_dynamic_spiders-1.0.0a1-py3-none-any.whl size=10526 sha256=f59967d02836c97b570bff1f3609d3bbca3e0408c1ce0df87621cbfab084240c Stored in directory: /builddir/.cache/pip/wheels/7a/25/d5/c1d72a53043eef3b984e4403debcec681b82c982cb901d2fda Successfully built scrapy-dynamic-spiders + RPM_EC=0 ++ jobs -p + exit 0 Executing(%install): /bin/sh -e /var/tmp/rpm-tmp.HfrPh3 + umask 022 + cd /builddir/build/BUILD + '[' /builddir/build/BUILDROOT/python-scrapy-dynamic-spiders-1.0.0~a1-1.fc37.x86_64 '!=' / ']' + rm -rf /builddir/build/BUILDROOT/python-scrapy-dynamic-spiders-1.0.0~a1-1.fc37.x86_64 ++ dirname /builddir/build/BUILDROOT/python-scrapy-dynamic-spiders-1.0.0~a1-1.fc37.x86_64 + mkdir -p /builddir/build/BUILDROOT + mkdir /builddir/build/BUILDROOT/python-scrapy-dynamic-spiders-1.0.0~a1-1.fc37.x86_64 + CFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection' + export CFLAGS + CXXFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection' + export CXXFLAGS + FFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -I/usr/lib64/gfortran/modules' + export FFLAGS + FCFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -I/usr/lib64/gfortran/modules' + export FCFLAGS + LDFLAGS='-Wl,-z,relro -Wl,--as-needed -Wl,-z,now -specs=/usr/lib/rpm/redhat/redhat-hardened-ld -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -Wl,--build-id=sha1 -Wl,-dT,/builddir/build/BUILD/scrapy_dynamic_spiders-1.0.0a1/.package_note-python-scrapy-dynamic-spiders-1.0.0~a1-1.fc37.x86_64.ld' + export LDFLAGS + LT_SYS_LIBRARY_PATH=/usr/lib64: + export LT_SYS_LIBRARY_PATH + CC=gcc + export CC + CXX=g++ + export CXX + cd scrapy_dynamic_spiders-1.0.0a1 ++ ls /builddir/build/BUILD/scrapy_dynamic_spiders-1.0.0a1/pyproject-wheeldir/scrapy_dynamic_spiders-1.0.0a1-py3-none-any.whl ++ sed -E 's/([^-]+)-([^-]+)-.+\.whl/\1==\2/' ++ xargs basename --multiple + specifier=scrapy_dynamic_spiders==1.0.0a1 + TMPDIR=/builddir/build/BUILD/scrapy_dynamic_spiders-1.0.0a1/.pyproject-builddir + /usr/bin/python3 -m pip install --root /builddir/build/BUILDROOT/python-scrapy-dynamic-spiders-1.0.0~a1-1.fc37.x86_64 --no-deps --disable-pip-version-check --progress-bar off --verbose --ignore-installed --no-warn-script-location --no-index --no-cache-dir --find-links /builddir/build/BUILD/scrapy_dynamic_spiders-1.0.0a1/pyproject-wheeldir scrapy_dynamic_spiders==1.0.0a1 Using pip 22.0.4 from /usr/lib/python3.10/site-packages/pip (python 3.10) Looking in links: /builddir/build/BUILD/scrapy_dynamic_spiders-1.0.0a1/pyproject-wheeldir Processing ./pyproject-wheeldir/scrapy_dynamic_spiders-1.0.0a1-py3-none-any.whl Installing collected packages: scrapy_dynamic_spiders Successfully installed scrapy_dynamic_spiders-1.0.0a1 + '[' -d /builddir/build/BUILDROOT/python-scrapy-dynamic-spiders-1.0.0~a1-1.fc37.x86_64/usr/bin ']' + rm -f /builddir/build/BUILD/python-scrapy-dynamic-spiders-1.0.0~a1-1.fc37.x86_64-pyproject-ghost-distinfo + site_dirs=() + '[' -d /builddir/build/BUILDROOT/python-scrapy-dynamic-spiders-1.0.0~a1-1.fc37.x86_64/usr/lib/python3.10/site-packages ']' + site_dirs+=("/usr/lib/python3.10/site-packages") + '[' /builddir/build/BUILDROOT/python-scrapy-dynamic-spiders-1.0.0~a1-1.fc37.x86_64/usr/lib64/python3.10/site-packages '!=' /builddir/build/BUILDROOT/python-scrapy-dynamic-spiders-1.0.0~a1-1.fc37.x86_64/usr/lib/python3.10/site-packages ']' + '[' -d /builddir/build/BUILDROOT/python-scrapy-dynamic-spiders-1.0.0~a1-1.fc37.x86_64/usr/lib64/python3.10/site-packages ']' + for site_dir in ${site_dirs[@]} + for distinfo in /builddir/build/BUILDROOT/python-scrapy-dynamic-spiders-1.0.0~a1-1.fc37.x86_64$site_dir/*.dist-info + echo '%ghost /usr/lib/python3.10/site-packages/scrapy_dynamic_spiders-1.0.0a1.dist-info' + sed -i s/pip/rpm/ /builddir/build/BUILDROOT/python-scrapy-dynamic-spiders-1.0.0~a1-1.fc37.x86_64/usr/lib/python3.10/site-packages/scrapy_dynamic_spiders-1.0.0a1.dist-info/INSTALLER + PYTHONPATH=/usr/lib/rpm/redhat + /usr/bin/python3 -B /usr/lib/rpm/redhat/pyproject_preprocess_record.py --buildroot /builddir/build/BUILDROOT/python-scrapy-dynamic-spiders-1.0.0~a1-1.fc37.x86_64 --record /builddir/build/BUILDROOT/python-scrapy-dynamic-spiders-1.0.0~a1-1.fc37.x86_64/usr/lib/python3.10/site-packages/scrapy_dynamic_spiders-1.0.0a1.dist-info/RECORD --output /builddir/build/BUILD/python-scrapy-dynamic-spiders-1.0.0~a1-1.fc37.x86_64-pyproject-record + rm -fv /builddir/build/BUILDROOT/python-scrapy-dynamic-spiders-1.0.0~a1-1.fc37.x86_64/usr/lib/python3.10/site-packages/scrapy_dynamic_spiders-1.0.0a1.dist-info/RECORD removed '/builddir/build/BUILDROOT/python-scrapy-dynamic-spiders-1.0.0~a1-1.fc37.x86_64/usr/lib/python3.10/site-packages/scrapy_dynamic_spiders-1.0.0a1.dist-info/RECORD' + rm -fv /builddir/build/BUILDROOT/python-scrapy-dynamic-spiders-1.0.0~a1-1.fc37.x86_64/usr/lib/python3.10/site-packages/scrapy_dynamic_spiders-1.0.0a1.dist-info/REQUESTED removed '/builddir/build/BUILDROOT/python-scrapy-dynamic-spiders-1.0.0~a1-1.fc37.x86_64/usr/lib/python3.10/site-packages/scrapy_dynamic_spiders-1.0.0a1.dist-info/REQUESTED' ++ wc -l /builddir/build/BUILD/python-scrapy-dynamic-spiders-1.0.0~a1-1.fc37.x86_64-pyproject-ghost-distinfo ++ cut -f1 '-d ' + lines=1 + '[' 1 -ne 1 ']' + /usr/bin/python3 /usr/lib/rpm/redhat/pyproject_save_files.py --output-files /builddir/build/BUILD/python-scrapy-dynamic-spiders-1.0.0~a1-1.fc37.x86_64-pyproject-files --output-modules /builddir/build/BUILD/python-scrapy-dynamic-spiders-1.0.0~a1-1.fc37.x86_64-pyproject-modules --buildroot /builddir/build/BUILDROOT/python-scrapy-dynamic-spiders-1.0.0~a1-1.fc37.x86_64 --sitelib /usr/lib/python3.10/site-packages --sitearch /usr/lib64/python3.10/site-packages --python-version 3.10 --pyproject-record /builddir/build/BUILD/python-scrapy-dynamic-spiders-1.0.0~a1-1.fc37.x86_64-pyproject-record --prefix /usr '*' +auto + /usr/bin/find-debuginfo -j2 --strict-build-id -m -i --build-id-seed 1.0.0~a1-1.fc37 --unique-debug-suffix -1.0.0~a1-1.fc37.x86_64 --unique-debug-src-base python-scrapy-dynamic-spiders-1.0.0~a1-1.fc37.x86_64 --run-dwz --dwz-low-mem-die-limit 10000000 --dwz-max-die-limit 110000000 -S debugsourcefiles.list /builddir/build/BUILD/scrapy_dynamic_spiders-1.0.0a1 find: 'debug': No such file or directory + /usr/lib/rpm/check-buildroot + /usr/lib/rpm/redhat/brp-ldconfig + /usr/lib/rpm/brp-compress + /usr/lib/rpm/redhat/brp-strip-lto /usr/bin/strip + /usr/lib/rpm/brp-strip-static-archive /usr/bin/strip + /usr/lib/rpm/check-rpaths + /usr/lib/rpm/redhat/brp-mangle-shebangs + /usr/lib/rpm/redhat/brp-python-bytecompile '' 1 0 Bytecompiling .py files below /builddir/build/BUILDROOT/python-scrapy-dynamic-spiders-1.0.0~a1-1.fc37.x86_64/usr/lib/python3.10 using python3.10 + /usr/lib/rpm/redhat/brp-python-hardlink Executing(%check): /bin/sh -e /var/tmp/rpm-tmp.AMfLtB + umask 022 + cd /builddir/build/BUILD + CFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection' + export CFLAGS + CXXFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection' + export CXXFLAGS + FFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -I/usr/lib64/gfortran/modules' + export FFLAGS + FCFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -I/usr/lib64/gfortran/modules' + export FCFLAGS + LDFLAGS='-Wl,-z,relro -Wl,--as-needed -Wl,-z,now -specs=/usr/lib/rpm/redhat/redhat-hardened-ld -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -Wl,--build-id=sha1 -Wl,-dT,/builddir/build/BUILD/scrapy_dynamic_spiders-1.0.0a1/.package_note-python-scrapy-dynamic-spiders-1.0.0~a1-1.fc37.x86_64.ld' + export LDFLAGS + LT_SYS_LIBRARY_PATH=/usr/lib64: + export LT_SYS_LIBRARY_PATH + CC=gcc + export CC + CXX=g++ + export CXX + '[' -f /usr/lib/rpm/generate-rpm-note.sh ']' + /usr/lib/rpm/generate-rpm-note.sh python-scrapy-dynamic-spiders 1.0.0~a1-1.fc37 x86_64 + cd scrapy_dynamic_spiders-1.0.0a1 + '[' '!' -f /builddir/build/BUILD/python-scrapy-dynamic-spiders-1.0.0~a1-1.fc37.x86_64-pyproject-modules ']' + PATH=/builddir/build/BUILDROOT/python-scrapy-dynamic-spiders-1.0.0~a1-1.fc37.x86_64/usr/bin:/builddir/.local/bin:/builddir/bin:/usr/bin:/bin:/usr/sbin:/sbin:/usr/local/sbin + PYTHONPATH=/builddir/build/BUILDROOT/python-scrapy-dynamic-spiders-1.0.0~a1-1.fc37.x86_64/usr/lib64/python3.10/site-packages:/builddir/build/BUILDROOT/python-scrapy-dynamic-spiders-1.0.0~a1-1.fc37.x86_64/usr/lib/python3.10/site-packages + _PYTHONSITE=/builddir/build/BUILDROOT/python-scrapy-dynamic-spiders-1.0.0~a1-1.fc37.x86_64/usr/lib64/python3.10/site-packages:/builddir/build/BUILDROOT/python-scrapy-dynamic-spiders-1.0.0~a1-1.fc37.x86_64/usr/lib/python3.10/site-packages + PYTHONDONTWRITEBYTECODE=1 + /usr/bin/python3 -s /usr/lib/rpm/redhat/import_all_modules.py -f /builddir/build/BUILD/python-scrapy-dynamic-spiders-1.0.0~a1-1.fc37.x86_64-pyproject-modules -t Check import: scrapy_dynamic_spiders Check import: tests + RPM_EC=0 ++ jobs -p + exit 0 Processing files: python3-scrapy-dynamic-spiders-1.0.0~a1-1.fc37.noarch Provides: python-scrapy-dynamic-spiders = 1.0.0~a1-1.fc37 python3-scrapy-dynamic-spiders = 1.0.0~a1-1.fc37 python3.10-scrapy-dynamic-spiders = 1.0.0~a1-1.fc37 python3.10dist(scrapy-dynamic-spiders) = 1~a1 python3dist(scrapy-dynamic-spiders) = 1~a1 Requires(rpmlib): rpmlib(CompressedFileNames) <= 3.0.4-1 rpmlib(FileDigests) <= 4.6.0-1 rpmlib(PartialHardlinkSets) <= 4.0.4-1 rpmlib(PayloadFilesHavePrefix) <= 4.0-1 Requires: python(abi) = 3.10 python3.10dist(crochet) python3.10dist(scrapy) Checking for unpackaged file(s): /usr/lib/rpm/check-files /builddir/build/BUILDROOT/python-scrapy-dynamic-spiders-1.0.0~a1-1.fc37.x86_64 Wrote: /builddir/build/SRPMS/python-scrapy-dynamic-spiders-1.0.0~a1-1.fc37.src.rpm Wrote: /builddir/build/RPMS/python3-scrapy-dynamic-spiders-1.0.0~a1-1.fc37.noarch.rpm Executing(%clean): /bin/sh -e /var/tmp/rpm-tmp.xAgMf2 + umask 022 + cd /builddir/build/BUILD + cd scrapy_dynamic_spiders-1.0.0a1 + /usr/bin/rm -rf /builddir/build/BUILDROOT/python-scrapy-dynamic-spiders-1.0.0~a1-1.fc37.x86_64 + RPM_EC=0 ++ jobs -p + exit 0 Child return code was: 0