Mock Version: 6.7 Mock Version: 6.7 Mock Version: 6.7 Using nspawn with args ['--capability=cap_ipc_lock', '--bind=/tmp/mock-resolv.qop9f7d9:/etc/resolv.conf', '--bind=/dev/btrfs-control', '--bind=/dev/mapper/control', '--bind=/dev/fuse', '--bind=/dev/loop-control', '--bind=/dev/loop0', '--bind=/dev/loop1', '--bind=/dev/loop2', '--bind=/dev/loop3', '--bind=/dev/loop4', '--bind=/dev/loop5', '--bind=/dev/loop6', '--bind=/dev/loop7', '--bind=/dev/loop8', '--bind=/dev/loop9', '--bind=/dev/loop10', '--bind=/dev/loop11'] Executing command: ['/usr/bin/systemd-nspawn', '-q', '-M', '7350c01979524e9cada3e139f0f092f2', '-D', '/var/lib/mock/openstack-epoxy-alma-9-build-94-5/root', '-a', '-u', 'mockbuild', '--capability=cap_ipc_lock', '--bind=/tmp/mock-resolv.qop9f7d9:/etc/resolv.conf', '--bind=/dev/btrfs-control', '--bind=/dev/mapper/control', '--bind=/dev/fuse', '--bind=/dev/loop-control', '--bind=/dev/loop0', '--bind=/dev/loop1', '--bind=/dev/loop2', '--bind=/dev/loop3', '--bind=/dev/loop4', '--bind=/dev/loop5', '--bind=/dev/loop6', '--bind=/dev/loop7', '--bind=/dev/loop8', '--bind=/dev/loop9', '--bind=/dev/loop10', '--bind=/dev/loop11', '--console=pipe', '--setenv=TERM=vt100', '--setenv=SHELL=/bin/bash', '--setenv=HOME=/builddir', '--setenv=HOSTNAME=mock', '--setenv=PATH=/usr/bin:/bin:/usr/sbin:/sbin', '--setenv=PROMPT_COMMAND=printf "\\033]0;\\007"', '--setenv=PS1= \\s-\\v\\$ ', '--setenv=LANG=C.UTF-8', '--resolv-conf=off', 'bash', '--login', '-c', '/usr/bin/rpmbuild -bs --noclean --target x86_64 --nodeps /builddir/build/SPECS/python-confluent-kafka.spec'] with env {'TERM': 'vt100', 'SHELL': '/bin/bash', 'HOME': '/builddir', 'HOSTNAME': 'mock', 'PATH': '/usr/bin:/bin:/usr/sbin:/sbin', 'PROMPT_COMMAND': 'printf "\\033]0;\\007"', 'PS1': ' \\s-\\v\\$ ', 'LANG': 'C.UTF-8', 'SYSTEMD_NSPAWN_TMPFS_TMP': '0', 'SYSTEMD_SECCOMP': '0'} and shell False Building target platforms: x86_64 Building for target x86_64 setting SOURCE_DATE_EPOCH=1776384000 Wrote: /builddir/build/SRPMS/python-confluent-kafka-2.12.0-1.el9.src.rpm Child return code was: 0 Using nspawn with args ['--capability=cap_ipc_lock', '--bind=/tmp/mock-resolv.qop9f7d9:/etc/resolv.conf', '--bind=/dev/btrfs-control', '--bind=/dev/mapper/control', '--bind=/dev/fuse', '--bind=/dev/loop-control', '--bind=/dev/loop0', '--bind=/dev/loop1', '--bind=/dev/loop2', '--bind=/dev/loop3', '--bind=/dev/loop4', '--bind=/dev/loop5', '--bind=/dev/loop6', '--bind=/dev/loop7', '--bind=/dev/loop8', '--bind=/dev/loop9', '--bind=/dev/loop10', '--bind=/dev/loop11'] Executing command: ['/usr/bin/systemd-nspawn', '-q', '-M', 'dd2dc442601f48b4876f457d57a61c03', '-D', '/var/lib/mock/openstack-epoxy-alma-9-build-94-5/root', '-a', '-u', 'mockbuild', '--capability=cap_ipc_lock', '--bind=/tmp/mock-resolv.qop9f7d9:/etc/resolv.conf', '--bind=/dev/btrfs-control', '--bind=/dev/mapper/control', '--bind=/dev/fuse', '--bind=/dev/loop-control', '--bind=/dev/loop0', '--bind=/dev/loop1', '--bind=/dev/loop2', '--bind=/dev/loop3', '--bind=/dev/loop4', '--bind=/dev/loop5', '--bind=/dev/loop6', '--bind=/dev/loop7', '--bind=/dev/loop8', '--bind=/dev/loop9', '--bind=/dev/loop10', '--bind=/dev/loop11', '--console=pipe', '--setenv=TERM=vt100', '--setenv=SHELL=/bin/bash', '--setenv=HOME=/builddir', '--setenv=HOSTNAME=mock', '--setenv=PATH=/usr/bin:/bin:/usr/sbin:/sbin', '--setenv=PROMPT_COMMAND=printf "\\033]0;\\007"', '--setenv=PS1= \\s-\\v\\$ ', '--setenv=LANG=C.UTF-8', '--resolv-conf=off', 'bash', '--login', '-c', '/usr/bin/rpmbuild -bs --noclean --target x86_64 --nodeps /builddir/build/SPECS/python-confluent-kafka.spec'] with env {'TERM': 'vt100', 'SHELL': '/bin/bash', 'HOME': '/builddir', 'HOSTNAME': 'mock', 'PATH': '/usr/bin:/bin:/usr/sbin:/sbin', 'PROMPT_COMMAND': 'printf "\\033]0;\\007"', 'PS1': ' \\s-\\v\\$ ', 'LANG': 'C.UTF-8', 'SYSTEMD_NSPAWN_TMPFS_TMP': '0', 'SYSTEMD_SECCOMP': '0'} and shell False Building target platforms: x86_64 Building for target x86_64 setting SOURCE_DATE_EPOCH=1776384000 Wrote: /builddir/build/SRPMS/python-confluent-kafka-2.12.0-1.el9.src.rpm Child return code was: 0 Using nspawn with args ['--capability=cap_ipc_lock', '--bind=/tmp/mock-resolv.qop9f7d9:/etc/resolv.conf', '--bind=/dev/btrfs-control', '--bind=/dev/mapper/control', '--bind=/dev/fuse', '--bind=/dev/loop-control', '--bind=/dev/loop0', '--bind=/dev/loop1', '--bind=/dev/loop2', '--bind=/dev/loop3', '--bind=/dev/loop4', '--bind=/dev/loop5', '--bind=/dev/loop6', '--bind=/dev/loop7', '--bind=/dev/loop8', '--bind=/dev/loop9', '--bind=/dev/loop10', '--bind=/dev/loop11'] Executing command: ['/usr/bin/systemd-nspawn', '-q', '-M', 'a17e3d090cb24a21ae0c418932df9e1d', '-D', '/var/lib/mock/openstack-epoxy-alma-9-build-94-5/root', '-a', '-u', 'mockbuild', '--capability=cap_ipc_lock', '--bind=/tmp/mock-resolv.qop9f7d9:/etc/resolv.conf', '--bind=/dev/btrfs-control', '--bind=/dev/mapper/control', '--bind=/dev/fuse', '--bind=/dev/loop-control', '--bind=/dev/loop0', '--bind=/dev/loop1', '--bind=/dev/loop2', '--bind=/dev/loop3', '--bind=/dev/loop4', '--bind=/dev/loop5', '--bind=/dev/loop6', '--bind=/dev/loop7', '--bind=/dev/loop8', '--bind=/dev/loop9', '--bind=/dev/loop10', '--bind=/dev/loop11', '--console=pipe', '--setenv=TERM=vt100', '--setenv=SHELL=/bin/bash', '--setenv=HOME=/builddir', '--setenv=HOSTNAME=mock', '--setenv=PATH=/usr/bin:/bin:/usr/sbin:/sbin', '--setenv=PROMPT_COMMAND=printf "\\033]0;\\007"', '--setenv=PS1= \\s-\\v\\$ ', '--setenv=LANG=C.UTF-8', '--resolv-conf=off', 'bash', '--login', '-c', '/usr/bin/rpmbuild -br --noclean --target x86_64 --nodeps /builddir/build/SPECS/python-confluent-kafka.spec'] with env {'TERM': 'vt100', 'SHELL': '/bin/bash', 'HOME': '/builddir', 'HOSTNAME': 'mock', 'PATH': '/usr/bin:/bin:/usr/sbin:/sbin', 'PROMPT_COMMAND': 'printf "\\033]0;\\007"', 'PS1': ' \\s-\\v\\$ ', 'LANG': 'C.UTF-8', 'SYSTEMD_NSPAWN_TMPFS_TMP': '0', 'SYSTEMD_SECCOMP': '0'} and shell False Building target platforms: x86_64 Building for target x86_64 setting SOURCE_DATE_EPOCH=1776384000 Executing(%prep): /bin/sh -e /var/tmp/rpm-tmp.EgS8VE + umask 022 + cd /builddir/build/BUILD + cd /builddir/build/BUILD + rm -rf confluent_kafka-2.12.0 + /usr/bin/gzip -dc /builddir/build/SOURCES/confluent_kafka-2.12.0.tar.gz + /usr/bin/tar -xof - + STATUS=0 + '[' 0 -ne 0 ']' + cd confluent_kafka-2.12.0 + /usr/bin/chmod -Rf a+rX,u+w,g-w,o-w . + RPM_EC=0 ++ jobs -p + exit 0 Executing(%generate_buildrequires): /bin/sh -e /var/tmp/rpm-tmp.HpZfLf + umask 022 + cd /builddir/build/BUILD + cd confluent_kafka-2.12.0 + echo pyproject-rpm-macros + exit 0 Wrote: /builddir/build/SRPMS/python-confluent-kafka-2.12.0-1.el9.buildreqs.nosrc.rpm Child return code was: 11 Dynamic buildrequires detected Going to install missing buildrequires. See root.log for details. Using nspawn with args ['--capability=cap_ipc_lock', '--bind=/tmp/mock-resolv.qop9f7d9:/etc/resolv.conf', '--bind=/dev/btrfs-control', '--bind=/dev/mapper/control', '--bind=/dev/fuse', '--bind=/dev/loop-control', '--bind=/dev/loop0', '--bind=/dev/loop1', '--bind=/dev/loop2', '--bind=/dev/loop3', '--bind=/dev/loop4', '--bind=/dev/loop5', '--bind=/dev/loop6', '--bind=/dev/loop7', '--bind=/dev/loop8', '--bind=/dev/loop9', '--bind=/dev/loop10', '--bind=/dev/loop11'] Executing command: ['/usr/bin/systemd-nspawn', '-q', '-M', 'b29626c1289849afbc0c1d8b7c729273', '-D', '/var/lib/mock/openstack-epoxy-alma-9-build-94-5/root', '-a', '-u', 'mockbuild', '--capability=cap_ipc_lock', '--bind=/tmp/mock-resolv.qop9f7d9:/etc/resolv.conf', '--bind=/dev/btrfs-control', '--bind=/dev/mapper/control', '--bind=/dev/fuse', '--bind=/dev/loop-control', '--bind=/dev/loop0', '--bind=/dev/loop1', '--bind=/dev/loop2', '--bind=/dev/loop3', '--bind=/dev/loop4', '--bind=/dev/loop5', '--bind=/dev/loop6', '--bind=/dev/loop7', '--bind=/dev/loop8', '--bind=/dev/loop9', '--bind=/dev/loop10', '--bind=/dev/loop11', '--console=pipe', '--setenv=TERM=vt100', '--setenv=SHELL=/bin/bash', '--setenv=HOME=/builddir', '--setenv=HOSTNAME=mock', '--setenv=PATH=/usr/bin:/bin:/usr/sbin:/sbin', '--setenv=PROMPT_COMMAND=printf "\\033]0;\\007"', '--setenv=PS1= \\s-\\v\\$ ', '--setenv=LANG=C.UTF-8', '--resolv-conf=off', 'bash', '--login', '-c', '/usr/bin/rpmbuild -br --noprep --noclean --target x86_64 --nodeps /builddir/build/SPECS/python-confluent-kafka.spec'] with env {'TERM': 'vt100', 'SHELL': '/bin/bash', 'HOME': '/builddir', 'HOSTNAME': 'mock', 'PATH': '/usr/bin:/bin:/usr/sbin:/sbin', 'PROMPT_COMMAND': 'printf "\\033]0;\\007"', 'PS1': ' \\s-\\v\\$ ', 'LANG': 'C.UTF-8', 'SYSTEMD_NSPAWN_TMPFS_TMP': '0', 'SYSTEMD_SECCOMP': '0'} and shell False Building target platforms: x86_64 Building for target x86_64 setting SOURCE_DATE_EPOCH=1776384000 Executing(%generate_buildrequires): /bin/sh -e /var/tmp/rpm-tmp.jiT9My + umask 022 + cd /builddir/build/BUILD + cd confluent_kafka-2.12.0 + echo pyproject-rpm-macros + echo python3-devel + echo 'python3dist(packaging)' + echo 'python3dist(pip) >= 19' + '[' -f pyproject.toml ']' + echo '(python3dist(tomli) if python3-devel < 3.11)' + rm -rfv '*.dist-info/' + '[' -f /usr/bin/python3 ']' + mkdir -p /builddir/build/BUILD/confluent_kafka-2.12.0/.pyproject-builddir + echo -n + CFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -march=x86-64-v2 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection' + CXXFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -march=x86-64-v2 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection' + FFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -march=x86-64-v2 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -I/usr/lib64/gfortran/modules' + FCFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -march=x86-64-v2 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -I/usr/lib64/gfortran/modules' + LDFLAGS='-Wl,-z,relro -Wl,--as-needed -Wl,-z,now -specs=/usr/lib/rpm/redhat/redhat-hardened-ld -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 ' + LT_SYS_LIBRARY_PATH=/usr/lib64: + CC=gcc + CXX=g++ + TMPDIR=/builddir/build/BUILD/confluent_kafka-2.12.0/.pyproject-builddir + RPM_TOXENV=py39 + HOSTNAME=rpmbuild + /usr/bin/python3 -Bs /usr/lib/rpm/redhat/pyproject_buildrequires.py --generate-extras --python3_pkgversion 3 --wheeldir /builddir/build/BUILD/confluent_kafka-2.12.0/pyproject-wheeldir --output /builddir/build/BUILD/python-confluent-kafka-2.12.0-1.el9.x86_64-pyproject-buildrequires Import error: No module named 'tomli' + cat /builddir/build/BUILD/python-confluent-kafka-2.12.0-1.el9.x86_64-pyproject-buildrequires + rm -rfv '*.dist-info/' + RPM_EC=0 ++ jobs -p + exit 0 Wrote: /builddir/build/SRPMS/python-confluent-kafka-2.12.0-1.el9.buildreqs.nosrc.rpm Child return code was: 11 Dynamic buildrequires detected Going to install missing buildrequires. See root.log for details. Using nspawn with args ['--capability=cap_ipc_lock', '--bind=/tmp/mock-resolv.qop9f7d9:/etc/resolv.conf', '--bind=/dev/btrfs-control', '--bind=/dev/mapper/control', '--bind=/dev/fuse', '--bind=/dev/loop-control', '--bind=/dev/loop0', '--bind=/dev/loop1', '--bind=/dev/loop2', '--bind=/dev/loop3', '--bind=/dev/loop4', '--bind=/dev/loop5', '--bind=/dev/loop6', '--bind=/dev/loop7', '--bind=/dev/loop8', '--bind=/dev/loop9', '--bind=/dev/loop10', '--bind=/dev/loop11'] Executing command: ['/usr/bin/systemd-nspawn', '-q', '-M', 'a5c13dc1268143fd8d78727ee0d31132', '-D', '/var/lib/mock/openstack-epoxy-alma-9-build-94-5/root', '-a', '-u', 'mockbuild', '--capability=cap_ipc_lock', '--bind=/tmp/mock-resolv.qop9f7d9:/etc/resolv.conf', '--bind=/dev/btrfs-control', '--bind=/dev/mapper/control', '--bind=/dev/fuse', '--bind=/dev/loop-control', '--bind=/dev/loop0', '--bind=/dev/loop1', '--bind=/dev/loop2', '--bind=/dev/loop3', '--bind=/dev/loop4', '--bind=/dev/loop5', '--bind=/dev/loop6', '--bind=/dev/loop7', '--bind=/dev/loop8', '--bind=/dev/loop9', '--bind=/dev/loop10', '--bind=/dev/loop11', '--console=pipe', '--setenv=TERM=vt100', '--setenv=SHELL=/bin/bash', '--setenv=HOME=/builddir', '--setenv=HOSTNAME=mock', '--setenv=PATH=/usr/bin:/bin:/usr/sbin:/sbin', '--setenv=PROMPT_COMMAND=printf "\\033]0;\\007"', '--setenv=PS1= \\s-\\v\\$ ', '--setenv=LANG=C.UTF-8', '--resolv-conf=off', 'bash', '--login', '-c', '/usr/bin/rpmbuild -br --noprep --noclean --target x86_64 --nodeps /builddir/build/SPECS/python-confluent-kafka.spec'] with env {'TERM': 'vt100', 'SHELL': '/bin/bash', 'HOME': '/builddir', 'HOSTNAME': 'mock', 'PATH': '/usr/bin:/bin:/usr/sbin:/sbin', 'PROMPT_COMMAND': 'printf "\\033]0;\\007"', 'PS1': ' \\s-\\v\\$ ', 'LANG': 'C.UTF-8', 'SYSTEMD_NSPAWN_TMPFS_TMP': '0', 'SYSTEMD_SECCOMP': '0'} and shell False Building target platforms: x86_64 Building for target x86_64 setting SOURCE_DATE_EPOCH=1776384000 Executing(%generate_buildrequires): /bin/sh -e /var/tmp/rpm-tmp.SwSdr5 + umask 022 + cd /builddir/build/BUILD + cd confluent_kafka-2.12.0 + echo pyproject-rpm-macros + echo python3-devel + echo 'python3dist(packaging)' + echo 'python3dist(pip) >= 19' + '[' -f pyproject.toml ']' + echo '(python3dist(tomli) if python3-devel < 3.11)' + rm -rfv '*.dist-info/' + '[' -f /usr/bin/python3 ']' + mkdir -p /builddir/build/BUILD/confluent_kafka-2.12.0/.pyproject-builddir + echo -n + CFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -march=x86-64-v2 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection' + CXXFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -march=x86-64-v2 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection' + FFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -march=x86-64-v2 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -I/usr/lib64/gfortran/modules' + FCFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -march=x86-64-v2 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -I/usr/lib64/gfortran/modules' + LDFLAGS='-Wl,-z,relro -Wl,--as-needed -Wl,-z,now -specs=/usr/lib/rpm/redhat/redhat-hardened-ld -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 ' + LT_SYS_LIBRARY_PATH=/usr/lib64: + CC=gcc + CXX=g++ + TMPDIR=/builddir/build/BUILD/confluent_kafka-2.12.0/.pyproject-builddir + RPM_TOXENV=py39 + HOSTNAME=rpmbuild + /usr/bin/python3 -Bs /usr/lib/rpm/redhat/pyproject_buildrequires.py --generate-extras --python3_pkgversion 3 --wheeldir /builddir/build/BUILD/confluent_kafka-2.12.0/pyproject-wheeldir --output /builddir/build/BUILD/python-confluent-kafka-2.12.0-1.el9.x86_64-pyproject-buildrequires Handling setuptools>=62 from build-system.requires Requirement satisfied: setuptools>=62 (installed: setuptools 65.5.1) Handling wheel from build-system.requires Requirement not satisfied: wheel Exiting dependency generation pass: build backend + cat /builddir/build/BUILD/python-confluent-kafka-2.12.0-1.el9.x86_64-pyproject-buildrequires + rm -rfv '*.dist-info/' + RPM_EC=0 ++ jobs -p + exit 0 Wrote: /builddir/build/SRPMS/python-confluent-kafka-2.12.0-1.el9.buildreqs.nosrc.rpm Child return code was: 11 Dynamic buildrequires detected Going to install missing buildrequires. See root.log for details. Using nspawn with args ['--capability=cap_ipc_lock', '--bind=/tmp/mock-resolv.qop9f7d9:/etc/resolv.conf', '--bind=/dev/btrfs-control', '--bind=/dev/mapper/control', '--bind=/dev/fuse', '--bind=/dev/loop-control', '--bind=/dev/loop0', '--bind=/dev/loop1', '--bind=/dev/loop2', '--bind=/dev/loop3', '--bind=/dev/loop4', '--bind=/dev/loop5', '--bind=/dev/loop6', '--bind=/dev/loop7', '--bind=/dev/loop8', '--bind=/dev/loop9', '--bind=/dev/loop10', '--bind=/dev/loop11'] Executing command: ['/usr/bin/systemd-nspawn', '-q', '-M', '6297db5e59f545e89e06df5f64f126fb', '-D', '/var/lib/mock/openstack-epoxy-alma-9-build-94-5/root', '-a', '-u', 'mockbuild', '--capability=cap_ipc_lock', '--bind=/tmp/mock-resolv.qop9f7d9:/etc/resolv.conf', '--bind=/dev/btrfs-control', '--bind=/dev/mapper/control', '--bind=/dev/fuse', '--bind=/dev/loop-control', '--bind=/dev/loop0', '--bind=/dev/loop1', '--bind=/dev/loop2', '--bind=/dev/loop3', '--bind=/dev/loop4', '--bind=/dev/loop5', '--bind=/dev/loop6', '--bind=/dev/loop7', '--bind=/dev/loop8', '--bind=/dev/loop9', '--bind=/dev/loop10', '--bind=/dev/loop11', '--console=pipe', '--setenv=TERM=vt100', '--setenv=SHELL=/bin/bash', '--setenv=HOME=/builddir', '--setenv=HOSTNAME=mock', '--setenv=PATH=/usr/bin:/bin:/usr/sbin:/sbin', '--setenv=PROMPT_COMMAND=printf "\\033]0;\\007"', '--setenv=PS1= \\s-\\v\\$ ', '--setenv=LANG=C.UTF-8', '--resolv-conf=off', 'bash', '--login', '-c', '/usr/bin/rpmbuild -br --noprep --noclean --target x86_64 --nodeps /builddir/build/SPECS/python-confluent-kafka.spec'] with env {'TERM': 'vt100', 'SHELL': '/bin/bash', 'HOME': '/builddir', 'HOSTNAME': 'mock', 'PATH': '/usr/bin:/bin:/usr/sbin:/sbin', 'PROMPT_COMMAND': 'printf "\\033]0;\\007"', 'PS1': ' \\s-\\v\\$ ', 'LANG': 'C.UTF-8', 'SYSTEMD_NSPAWN_TMPFS_TMP': '0', 'SYSTEMD_SECCOMP': '0'} and shell False Building target platforms: x86_64 Building for target x86_64 setting SOURCE_DATE_EPOCH=1776384000 Executing(%generate_buildrequires): /bin/sh -e /var/tmp/rpm-tmp.yvX7sw + umask 022 + cd /builddir/build/BUILD + cd confluent_kafka-2.12.0 + echo pyproject-rpm-macros + echo python3-devel + echo 'python3dist(packaging)' + echo 'python3dist(pip) >= 19' + '[' -f pyproject.toml ']' + echo '(python3dist(tomli) if python3-devel < 3.11)' + rm -rfv '*.dist-info/' + '[' -f /usr/bin/python3 ']' + mkdir -p /builddir/build/BUILD/confluent_kafka-2.12.0/.pyproject-builddir + echo -n + CFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -march=x86-64-v2 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection' + CXXFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -march=x86-64-v2 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection' + FFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -march=x86-64-v2 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -I/usr/lib64/gfortran/modules' + FCFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -march=x86-64-v2 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -I/usr/lib64/gfortran/modules' + LDFLAGS='-Wl,-z,relro -Wl,--as-needed -Wl,-z,now -specs=/usr/lib/rpm/redhat/redhat-hardened-ld -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 ' + LT_SYS_LIBRARY_PATH=/usr/lib64: + CC=gcc + CXX=g++ + TMPDIR=/builddir/build/BUILD/confluent_kafka-2.12.0/.pyproject-builddir + RPM_TOXENV=py39 + HOSTNAME=rpmbuild + /usr/bin/python3 -Bs /usr/lib/rpm/redhat/pyproject_buildrequires.py --generate-extras --python3_pkgversion 3 --wheeldir /builddir/build/BUILD/confluent_kafka-2.12.0/pyproject-wheeldir --output /builddir/build/BUILD/python-confluent-kafka-2.12.0-1.el9.x86_64-pyproject-buildrequires Handling setuptools>=62 from build-system.requires Requirement satisfied: setuptools>=62 (installed: setuptools 65.5.1) Handling wheel from build-system.requires Requirement satisfied: wheel (installed: wheel 0.36.2) /usr/lib/python3.9/site-packages/setuptools/config/pyprojecttoml.py:108: _BetaConfiguration: Support for `[tool.setuptools]` in `pyproject.toml` is still *beta*. warnings.warn(msg, _BetaConfiguration) running egg_info writing src/confluent_kafka.egg-info/PKG-INFO writing dependency_links to src/confluent_kafka.egg-info/dependency_links.txt writing requirements to src/confluent_kafka.egg-info/requires.txt writing top-level names to src/confluent_kafka.egg-info/top_level.txt reading manifest file 'src/confluent_kafka.egg-info/SOURCES.txt' reading manifest template 'MANIFEST.in' no previously-included directories found matching 'tests' no previously-included directories found matching 'docs' adding license file 'LICENSE' writing manifest file 'src/confluent_kafka.egg-info/SOURCES.txt' Handling wheel from get_requires_for_build_wheel Requirement satisfied: wheel (installed: wheel 0.36.2) /usr/lib/python3.9/site-packages/setuptools/config/pyprojecttoml.py:108: _BetaConfiguration: Support for `[tool.setuptools]` in `pyproject.toml` is still *beta*. warnings.warn(msg, _BetaConfiguration) running dist_info creating confluent_kafka.egg-info writing confluent_kafka.egg-info/PKG-INFO writing dependency_links to confluent_kafka.egg-info/dependency_links.txt writing requirements to confluent_kafka.egg-info/requires.txt writing top-level names to confluent_kafka.egg-info/top_level.txt writing manifest file 'confluent_kafka.egg-info/SOURCES.txt' reading manifest file 'confluent_kafka.egg-info/SOURCES.txt' reading manifest template 'MANIFEST.in' no previously-included directories found matching 'tests' no previously-included directories found matching 'docs' adding license file 'LICENSE' writing manifest file 'confluent_kafka.egg-info/SOURCES.txt' creating '/builddir/build/BUILD/confluent_kafka-2.12.0/confluent_kafka-2.12.0.dist-info' adding license file "LICENSE" (matched pattern "LICEN[CS]E*") Handling tomli ; python_version < "3.13" from hook generated metadata: Requires-Dist (confluent-kafka) Requirement satisfied: tomli ; python_version < "3.13" (installed: tomli 2.0.1) Handling psutil ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: psutil ; extra == 'all' Handling opentelemetry-distro ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: opentelemetry-distro ; extra == 'all' Handling opentelemetry-exporter-otlp ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: opentelemetry-exporter-otlp ; extra == 'all' Handling sphinx ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: sphinx ; extra == 'all' Handling sphinx-rtd-theme ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: sphinx-rtd-theme ; extra == 'all' Handling pandoc ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: pandoc ; extra == 'all' Handling confluent-kafka ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: confluent-kafka ; extra == 'all' Handling fastapi ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: fastapi ; extra == 'all' Handling pydantic ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: pydantic ; extra == 'all' Handling uvicorn ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: uvicorn ; extra == 'all' Handling six ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: six ; extra == 'all' Handling attrs ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: attrs ; extra == 'all' Handling cachetools ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: cachetools ; extra == 'all' Handling httpx (>=0.26) ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: httpx (>=0.26) ; extra == 'all' Handling authlib (>=1.0.0) ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: authlib (>=1.0.0) ; extra == 'all' Handling requests ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: requests ; extra == 'all' Handling avro (<2,>=1.11.1) ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: avro (<2,>=1.11.1) ; extra == 'all' Handling pyrsistent ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: pyrsistent ; extra == 'all' Handling jsonschema ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: jsonschema ; extra == 'all' Handling orjson (>=3.10) ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: orjson (>=3.10) ; extra == 'all' Handling googleapis-common-protos ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: googleapis-common-protos ; extra == 'all' Handling protobuf ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: protobuf ; extra == 'all' Handling azure-identity ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: azure-identity ; extra == 'all' Handling azure-keyvault-keys ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: azure-keyvault-keys ; extra == 'all' Handling boto3 ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: boto3 ; extra == 'all' Handling cel-python (>=0.4.0) ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: cel-python (>=0.4.0) ; extra == 'all' Handling google-auth ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: google-auth ; extra == 'all' Handling google-api-core ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: google-api-core ; extra == 'all' Handling google-cloud-kms ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: google-cloud-kms ; extra == 'all' Handling hkdf (==0.0.3) ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: hkdf (==0.0.3) ; extra == 'all' Handling hvac ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: hvac ; extra == 'all' Handling jsonata-python ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: jsonata-python ; extra == 'all' Handling pyyaml (>=6.0.0) ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: pyyaml (>=6.0.0) ; extra == 'all' Handling tink ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: tink ; extra == 'all' Handling urllib3 (<3) ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: urllib3 (<3) ; extra == 'all' Handling flake8 ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: flake8 ; extra == 'all' Handling orjson ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: orjson ; extra == 'all' Handling pytest ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: pytest ; extra == 'all' Handling pytest-timeout ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: pytest-timeout ; extra == 'all' Handling requests-mock ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: requests-mock ; extra == 'all' Handling respx ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: respx ; extra == 'all' Handling pytest-cov ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: pytest-cov ; extra == 'all' Handling pluggy (<1.6.0) ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: pluggy (<1.6.0) ; extra == 'all' Handling pytest-asyncio ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: pytest-asyncio ; extra == 'all' Handling async-timeout ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: async-timeout ; extra == 'all' Handling attrs (>=21.2.0) ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: attrs (>=21.2.0) ; extra == 'all' Handling cachetools (>=5.5.0) ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: cachetools (>=5.5.0) ; extra == 'all' Handling certifi ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: certifi ; extra == 'all' Handling boto3 (>=1.35) ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: boto3 (>=1.35) ; extra == 'all' Handling tomli ; (python_version < "3.13") and extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: tomli ; (python_version < "3.13") and extra == 'all' Handling fastavro (<1.8.0) ; (python_version == "3.7") and extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: fastavro (<1.8.0) ; (python_version == "3.7") and extra == 'all' Handling fastavro (<2) ; (python_version > "3.7") and extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: fastavro (<2) ; (python_version > "3.7") and extra == 'all' Handling requests ; extra == 'avro' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: requests ; extra == 'avro' Handling avro (<2,>=1.11.1) ; extra == 'avro' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: avro (<2,>=1.11.1) ; extra == 'avro' Handling attrs (>=21.2.0) ; extra == 'avro' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: attrs (>=21.2.0) ; extra == 'avro' Handling cachetools (>=5.5.0) ; extra == 'avro' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: cachetools (>=5.5.0) ; extra == 'avro' Handling certifi ; extra == 'avro' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: certifi ; extra == 'avro' Handling httpx (>=0.26) ; extra == 'avro' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: httpx (>=0.26) ; extra == 'avro' Handling authlib (>=1.0.0) ; extra == 'avro' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: authlib (>=1.0.0) ; extra == 'avro' Handling fastavro (<1.8.0) ; (python_version == "3.7") and extra == 'avro' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: fastavro (<1.8.0) ; (python_version == "3.7") and extra == 'avro' Handling fastavro (<2) ; (python_version > "3.7") and extra == 'avro' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: fastavro (<2) ; (python_version > "3.7") and extra == 'avro' Handling sphinx ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: sphinx ; extra == 'dev' Handling sphinx-rtd-theme ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: sphinx-rtd-theme ; extra == 'dev' Handling pandoc ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: pandoc ; extra == 'dev' Handling confluent-kafka ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: confluent-kafka ; extra == 'dev' Handling fastapi ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: fastapi ; extra == 'dev' Handling pydantic ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: pydantic ; extra == 'dev' Handling uvicorn ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: uvicorn ; extra == 'dev' Handling six ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: six ; extra == 'dev' Handling attrs ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: attrs ; extra == 'dev' Handling cachetools ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: cachetools ; extra == 'dev' Handling httpx (>=0.26) ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: httpx (>=0.26) ; extra == 'dev' Handling authlib (>=1.0.0) ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: authlib (>=1.0.0) ; extra == 'dev' Handling requests ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: requests ; extra == 'dev' Handling avro (<2,>=1.11.1) ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: avro (<2,>=1.11.1) ; extra == 'dev' Handling pyrsistent ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: pyrsistent ; extra == 'dev' Handling jsonschema ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: jsonschema ; extra == 'dev' Handling orjson (>=3.10) ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: orjson (>=3.10) ; extra == 'dev' Handling googleapis-common-protos ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: googleapis-common-protos ; extra == 'dev' Handling protobuf ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: protobuf ; extra == 'dev' Handling azure-identity ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: azure-identity ; extra == 'dev' Handling azure-keyvault-keys ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: azure-keyvault-keys ; extra == 'dev' Handling boto3 ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: boto3 ; extra == 'dev' Handling cel-python (>=0.4.0) ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: cel-python (>=0.4.0) ; extra == 'dev' Handling google-auth ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: google-auth ; extra == 'dev' Handling google-api-core ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: google-api-core ; extra == 'dev' Handling google-cloud-kms ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: google-cloud-kms ; extra == 'dev' Handling hkdf (==0.0.3) ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: hkdf (==0.0.3) ; extra == 'dev' Handling hvac ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: hvac ; extra == 'dev' Handling jsonata-python ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: jsonata-python ; extra == 'dev' Handling pyyaml (>=6.0.0) ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: pyyaml (>=6.0.0) ; extra == 'dev' Handling tink ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: tink ; extra == 'dev' Handling urllib3 (<3) ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: urllib3 (<3) ; extra == 'dev' Handling flake8 ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: flake8 ; extra == 'dev' Handling orjson ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: orjson ; extra == 'dev' Handling pytest ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: pytest ; extra == 'dev' Handling pytest-timeout ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: pytest-timeout ; extra == 'dev' Handling requests-mock ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: requests-mock ; extra == 'dev' Handling respx ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: respx ; extra == 'dev' Handling pytest-cov ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: pytest-cov ; extra == 'dev' Handling pluggy (<1.6.0) ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: pluggy (<1.6.0) ; extra == 'dev' Handling pytest-asyncio ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: pytest-asyncio ; extra == 'dev' Handling async-timeout ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: async-timeout ; extra == 'dev' Handling attrs (>=21.2.0) ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: attrs (>=21.2.0) ; extra == 'dev' Handling cachetools (>=5.5.0) ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: cachetools (>=5.5.0) ; extra == 'dev' Handling certifi ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: certifi ; extra == 'dev' Handling boto3 (>=1.35) ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: boto3 (>=1.35) ; extra == 'dev' Handling tomli ; (python_version < "3.13") and extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: tomli ; (python_version < "3.13") and extra == 'dev' Handling fastavro (<1.8.0) ; (python_version == "3.7") and extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: fastavro (<1.8.0) ; (python_version == "3.7") and extra == 'dev' Handling fastavro (<2) ; (python_version > "3.7") and extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: fastavro (<2) ; (python_version > "3.7") and extra == 'dev' Handling sphinx ; extra == 'docs' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: sphinx ; extra == 'docs' Handling sphinx-rtd-theme ; extra == 'docs' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: sphinx-rtd-theme ; extra == 'docs' Handling pandoc ; extra == 'docs' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: pandoc ; extra == 'docs' Handling attrs (>=21.2.0) ; extra == 'docs' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: attrs (>=21.2.0) ; extra == 'docs' Handling cachetools (>=5.5.0) ; extra == 'docs' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: cachetools (>=5.5.0) ; extra == 'docs' Handling certifi ; extra == 'docs' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: certifi ; extra == 'docs' Handling httpx (>=0.26) ; extra == 'docs' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: httpx (>=0.26) ; extra == 'docs' Handling authlib (>=1.0.0) ; extra == 'docs' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: authlib (>=1.0.0) ; extra == 'docs' Handling azure-identity ; extra == 'docs' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: azure-identity ; extra == 'docs' Handling azure-keyvault-keys ; extra == 'docs' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: azure-keyvault-keys ; extra == 'docs' Handling boto3 (>=1.35) ; extra == 'docs' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: boto3 (>=1.35) ; extra == 'docs' Handling cel-python (>=0.4.0) ; extra == 'docs' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: cel-python (>=0.4.0) ; extra == 'docs' Handling google-auth ; extra == 'docs' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: google-auth ; extra == 'docs' Handling google-api-core ; extra == 'docs' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: google-api-core ; extra == 'docs' Handling google-cloud-kms ; extra == 'docs' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: google-cloud-kms ; extra == 'docs' Handling hkdf (==0.0.3) ; extra == 'docs' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: hkdf (==0.0.3) ; extra == 'docs' Handling hvac ; extra == 'docs' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: hvac ; extra == 'docs' Handling jsonata-python ; extra == 'docs' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: jsonata-python ; extra == 'docs' Handling pyyaml (>=6.0.0) ; extra == 'docs' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: pyyaml (>=6.0.0) ; extra == 'docs' Handling tink ; extra == 'docs' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: tink ; extra == 'docs' Handling requests ; extra == 'docs' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: requests ; extra == 'docs' Handling avro (<2,>=1.11.1) ; extra == 'docs' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: avro (<2,>=1.11.1) ; extra == 'docs' Handling pyrsistent ; extra == 'docs' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: pyrsistent ; extra == 'docs' Handling jsonschema ; extra == 'docs' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: jsonschema ; extra == 'docs' Handling orjson (>=3.10) ; extra == 'docs' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: orjson (>=3.10) ; extra == 'docs' Handling googleapis-common-protos ; extra == 'docs' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: googleapis-common-protos ; extra == 'docs' Handling protobuf ; extra == 'docs' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: protobuf ; extra == 'docs' Handling tomli ; (python_version < "3.13") and extra == 'docs' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: tomli ; (python_version < "3.13") and extra == 'docs' Handling fastavro (<1.8.0) ; (python_version == "3.7") and extra == 'docs' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: fastavro (<1.8.0) ; (python_version == "3.7") and extra == 'docs' Handling fastavro (<2) ; (python_version > "3.7") and extra == 'docs' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: fastavro (<2) ; (python_version > "3.7") and extra == 'docs' Handling confluent-kafka ; extra == 'examples' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: confluent-kafka ; extra == 'examples' Handling fastapi ; extra == 'examples' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: fastapi ; extra == 'examples' Handling pydantic ; extra == 'examples' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: pydantic ; extra == 'examples' Handling uvicorn ; extra == 'examples' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: uvicorn ; extra == 'examples' Handling six ; extra == 'examples' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: six ; extra == 'examples' Handling attrs ; extra == 'examples' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: attrs ; extra == 'examples' Handling cachetools ; extra == 'examples' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: cachetools ; extra == 'examples' Handling httpx (>=0.26) ; extra == 'examples' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: httpx (>=0.26) ; extra == 'examples' Handling authlib (>=1.0.0) ; extra == 'examples' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: authlib (>=1.0.0) ; extra == 'examples' Handling requests ; extra == 'examples' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: requests ; extra == 'examples' Handling avro (<2,>=1.11.1) ; extra == 'examples' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: avro (<2,>=1.11.1) ; extra == 'examples' Handling pyrsistent ; extra == 'examples' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: pyrsistent ; extra == 'examples' Handling jsonschema ; extra == 'examples' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: jsonschema ; extra == 'examples' Handling orjson (>=3.10) ; extra == 'examples' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: orjson (>=3.10) ; extra == 'examples' Handling googleapis-common-protos ; extra == 'examples' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: googleapis-common-protos ; extra == 'examples' Handling protobuf ; extra == 'examples' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: protobuf ; extra == 'examples' Handling azure-identity ; extra == 'examples' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: azure-identity ; extra == 'examples' Handling azure-keyvault-keys ; extra == 'examples' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: azure-keyvault-keys ; extra == 'examples' Handling boto3 ; extra == 'examples' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: boto3 ; extra == 'examples' Handling cel-python (>=0.4.0) ; extra == 'examples' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: cel-python (>=0.4.0) ; extra == 'examples' Handling google-auth ; extra == 'examples' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: google-auth ; extra == 'examples' Handling google-api-core ; extra == 'examples' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: google-api-core ; extra == 'examples' Handling google-cloud-kms ; extra == 'examples' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: google-cloud-kms ; extra == 'examples' Handling hkdf (==0.0.3) ; extra == 'examples' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: hkdf (==0.0.3) ; extra == 'examples' Handling hvac ; extra == 'examples' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: hvac ; extra == 'examples' Handling jsonata-python ; extra == 'examples' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: jsonata-python ; extra == 'examples' Handling pyyaml (>=6.0.0) ; extra == 'examples' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: pyyaml (>=6.0.0) ; extra == 'examples' Handling tink ; extra == 'examples' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: tink ; extra == 'examples' Handling fastavro (<1.8.0) ; (python_version == "3.7") and extra == 'examples' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: fastavro (<1.8.0) ; (python_version == "3.7") and extra == 'examples' Handling fastavro (<2) ; (python_version > "3.7") and extra == 'examples' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: fastavro (<2) ; (python_version > "3.7") and extra == 'examples' Handling pyrsistent ; extra == 'json' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: pyrsistent ; extra == 'json' Handling jsonschema ; extra == 'json' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: jsonschema ; extra == 'json' Handling orjson (>=3.10) ; extra == 'json' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: orjson (>=3.10) ; extra == 'json' Handling attrs (>=21.2.0) ; extra == 'json' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: attrs (>=21.2.0) ; extra == 'json' Handling cachetools (>=5.5.0) ; extra == 'json' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: cachetools (>=5.5.0) ; extra == 'json' Handling certifi ; extra == 'json' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: certifi ; extra == 'json' Handling httpx (>=0.26) ; extra == 'json' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: httpx (>=0.26) ; extra == 'json' Handling authlib (>=1.0.0) ; extra == 'json' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: authlib (>=1.0.0) ; extra == 'json' Handling googleapis-common-protos ; extra == 'protobuf' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: googleapis-common-protos ; extra == 'protobuf' Handling protobuf ; extra == 'protobuf' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: protobuf ; extra == 'protobuf' Handling attrs (>=21.2.0) ; extra == 'protobuf' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: attrs (>=21.2.0) ; extra == 'protobuf' Handling cachetools (>=5.5.0) ; extra == 'protobuf' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: cachetools (>=5.5.0) ; extra == 'protobuf' Handling certifi ; extra == 'protobuf' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: certifi ; extra == 'protobuf' Handling httpx (>=0.26) ; extra == 'protobuf' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: httpx (>=0.26) ; extra == 'protobuf' Handling authlib (>=1.0.0) ; extra == 'protobuf' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: authlib (>=1.0.0) ; extra == 'protobuf' Handling azure-identity ; extra == 'rules' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: azure-identity ; extra == 'rules' Handling azure-keyvault-keys ; extra == 'rules' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: azure-keyvault-keys ; extra == 'rules' Handling boto3 (>=1.35) ; extra == 'rules' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: boto3 (>=1.35) ; extra == 'rules' Handling cel-python (>=0.4.0) ; extra == 'rules' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: cel-python (>=0.4.0) ; extra == 'rules' Handling google-auth ; extra == 'rules' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: google-auth ; extra == 'rules' Handling google-api-core ; extra == 'rules' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: google-api-core ; extra == 'rules' Handling google-cloud-kms ; extra == 'rules' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: google-cloud-kms ; extra == 'rules' Handling hkdf (==0.0.3) ; extra == 'rules' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: hkdf (==0.0.3) ; extra == 'rules' Handling hvac ; extra == 'rules' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: hvac ; extra == 'rules' Handling jsonata-python ; extra == 'rules' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: jsonata-python ; extra == 'rules' Handling pyyaml (>=6.0.0) ; extra == 'rules' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: pyyaml (>=6.0.0) ; extra == 'rules' Handling tink ; extra == 'rules' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: tink ; extra == 'rules' Handling attrs (>=21.2.0) ; extra == 'rules' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: attrs (>=21.2.0) ; extra == 'rules' Handling cachetools (>=5.5.0) ; extra == 'rules' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: cachetools (>=5.5.0) ; extra == 'rules' Handling certifi ; extra == 'rules' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: certifi ; extra == 'rules' Handling httpx (>=0.26) ; extra == 'rules' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: httpx (>=0.26) ; extra == 'rules' Handling authlib (>=1.0.0) ; extra == 'rules' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: authlib (>=1.0.0) ; extra == 'rules' Handling attrs (>=21.2.0) ; extra == 'schema_registry' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: attrs (>=21.2.0) ; extra == 'schema_registry' Handling cachetools (>=5.5.0) ; extra == 'schema_registry' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: cachetools (>=5.5.0) ; extra == 'schema_registry' Handling certifi ; extra == 'schema_registry' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: certifi ; extra == 'schema_registry' Handling httpx (>=0.26) ; extra == 'schema_registry' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: httpx (>=0.26) ; extra == 'schema_registry' Handling authlib (>=1.0.0) ; extra == 'schema_registry' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: authlib (>=1.0.0) ; extra == 'schema_registry' Handling attrs (>=21.2.0) ; extra == 'schemaregistry' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: attrs (>=21.2.0) ; extra == 'schemaregistry' Handling cachetools (>=5.5.0) ; extra == 'schemaregistry' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: cachetools (>=5.5.0) ; extra == 'schemaregistry' Handling certifi ; extra == 'schemaregistry' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: certifi ; extra == 'schemaregistry' Handling httpx (>=0.26) ; extra == 'schemaregistry' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: httpx (>=0.26) ; extra == 'schemaregistry' Handling authlib (>=1.0.0) ; extra == 'schemaregistry' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: authlib (>=1.0.0) ; extra == 'schemaregistry' Handling psutil ; extra == 'soaktest' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: psutil ; extra == 'soaktest' Handling opentelemetry-distro ; extra == 'soaktest' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: opentelemetry-distro ; extra == 'soaktest' Handling opentelemetry-exporter-otlp ; extra == 'soaktest' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: opentelemetry-exporter-otlp ; extra == 'soaktest' Handling urllib3 (<3) ; extra == 'tests' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: urllib3 (<3) ; extra == 'tests' Handling flake8 ; extra == 'tests' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: flake8 ; extra == 'tests' Handling orjson ; extra == 'tests' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: orjson ; extra == 'tests' Handling pytest ; extra == 'tests' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: pytest ; extra == 'tests' Handling pytest-timeout ; extra == 'tests' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: pytest-timeout ; extra == 'tests' Handling requests-mock ; extra == 'tests' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: requests-mock ; extra == 'tests' Handling respx ; extra == 'tests' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: respx ; extra == 'tests' Handling pytest-cov ; extra == 'tests' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: pytest-cov ; extra == 'tests' Handling pluggy (<1.6.0) ; extra == 'tests' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: pluggy (<1.6.0) ; extra == 'tests' Handling pytest-asyncio ; extra == 'tests' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: pytest-asyncio ; extra == 'tests' Handling async-timeout ; extra == 'tests' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: async-timeout ; extra == 'tests' Handling attrs (>=21.2.0) ; extra == 'tests' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: attrs (>=21.2.0) ; extra == 'tests' Handling cachetools (>=5.5.0) ; extra == 'tests' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: cachetools (>=5.5.0) ; extra == 'tests' Handling certifi ; extra == 'tests' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: certifi ; extra == 'tests' Handling httpx (>=0.26) ; extra == 'tests' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: httpx (>=0.26) ; extra == 'tests' Handling authlib (>=1.0.0) ; extra == 'tests' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: authlib (>=1.0.0) ; extra == 'tests' Handling azure-identity ; extra == 'tests' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: azure-identity ; extra == 'tests' Handling azure-keyvault-keys ; extra == 'tests' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: azure-keyvault-keys ; extra == 'tests' Handling boto3 (>=1.35) ; extra == 'tests' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: boto3 (>=1.35) ; extra == 'tests' Handling cel-python (>=0.4.0) ; extra == 'tests' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: cel-python (>=0.4.0) ; extra == 'tests' Handling google-auth ; extra == 'tests' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: google-auth ; extra == 'tests' Handling google-api-core ; extra == 'tests' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: google-api-core ; extra == 'tests' Handling google-cloud-kms ; extra == 'tests' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: google-cloud-kms ; extra == 'tests' Handling hkdf (==0.0.3) ; extra == 'tests' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: hkdf (==0.0.3) ; extra == 'tests' Handling hvac ; extra == 'tests' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: hvac ; extra == 'tests' Handling jsonata-python ; extra == 'tests' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: jsonata-python ; extra == 'tests' Handling pyyaml (>=6.0.0) ; extra == 'tests' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: pyyaml (>=6.0.0) ; extra == 'tests' Handling tink ; extra == 'tests' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: tink ; extra == 'tests' Handling requests ; extra == 'tests' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: requests ; extra == 'tests' Handling avro (<2,>=1.11.1) ; extra == 'tests' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: avro (<2,>=1.11.1) ; extra == 'tests' Handling pyrsistent ; extra == 'tests' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: pyrsistent ; extra == 'tests' Handling jsonschema ; extra == 'tests' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: jsonschema ; extra == 'tests' Handling orjson (>=3.10) ; extra == 'tests' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: orjson (>=3.10) ; extra == 'tests' Handling googleapis-common-protos ; extra == 'tests' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: googleapis-common-protos ; extra == 'tests' Handling protobuf ; extra == 'tests' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: protobuf ; extra == 'tests' Handling fastavro (<1.8.0) ; (python_version == "3.7") and extra == 'tests' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: fastavro (<1.8.0) ; (python_version == "3.7") and extra == 'tests' Handling fastavro (<2) ; (python_version > "3.7") and extra == 'tests' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: fastavro (<2) ; (python_version > "3.7") and extra == 'tests' + cat /builddir/build/BUILD/python-confluent-kafka-2.12.0-1.el9.x86_64-pyproject-buildrequires + rm -rfv confluent_kafka-2.12.0.dist-info/ removed 'confluent_kafka-2.12.0.dist-info/top_level.txt' removed 'confluent_kafka-2.12.0.dist-info/METADATA' removed 'confluent_kafka-2.12.0.dist-info/LICENSE' removed directory 'confluent_kafka-2.12.0.dist-info/' + RPM_EC=0 ++ jobs -p + exit 0 Wrote: /builddir/build/SRPMS/python-confluent-kafka-2.12.0-1.el9.buildreqs.nosrc.rpm Child return code was: 11 Dynamic buildrequires detected Going to install missing buildrequires. See root.log for details. Using nspawn with args ['--capability=cap_ipc_lock', '--bind=/tmp/mock-resolv.qop9f7d9:/etc/resolv.conf', '--bind=/dev/btrfs-control', '--bind=/dev/mapper/control', '--bind=/dev/fuse', '--bind=/dev/loop-control', '--bind=/dev/loop0', '--bind=/dev/loop1', '--bind=/dev/loop2', '--bind=/dev/loop3', '--bind=/dev/loop4', '--bind=/dev/loop5', '--bind=/dev/loop6', '--bind=/dev/loop7', '--bind=/dev/loop8', '--bind=/dev/loop9', '--bind=/dev/loop10', '--bind=/dev/loop11'] Executing command: ['/usr/bin/systemd-nspawn', '-q', '-M', '8af99f3236f74ffa84f1d0ce233af84e', '-D', '/var/lib/mock/openstack-epoxy-alma-9-build-94-5/root', '-a', '-u', 'mockbuild', '--capability=cap_ipc_lock', '--bind=/tmp/mock-resolv.qop9f7d9:/etc/resolv.conf', '--bind=/dev/btrfs-control', '--bind=/dev/mapper/control', '--bind=/dev/fuse', '--bind=/dev/loop-control', '--bind=/dev/loop0', '--bind=/dev/loop1', '--bind=/dev/loop2', '--bind=/dev/loop3', '--bind=/dev/loop4', '--bind=/dev/loop5', '--bind=/dev/loop6', '--bind=/dev/loop7', '--bind=/dev/loop8', '--bind=/dev/loop9', '--bind=/dev/loop10', '--bind=/dev/loop11', '--console=pipe', '--setenv=TERM=vt100', '--setenv=SHELL=/bin/bash', '--setenv=HOME=/builddir', '--setenv=HOSTNAME=mock', '--setenv=PATH=/usr/bin:/bin:/usr/sbin:/sbin', '--setenv=PROMPT_COMMAND=printf "\\033]0;\\007"', '--setenv=PS1= \\s-\\v\\$ ', '--setenv=LANG=C.UTF-8', '--resolv-conf=off', 'bash', '--login', '-c', '/usr/bin/rpmbuild -ba --noprep --noclean --target x86_64 /builddir/build/SPECS/python-confluent-kafka.spec'] with env {'TERM': 'vt100', 'SHELL': '/bin/bash', 'HOME': '/builddir', 'HOSTNAME': 'mock', 'PATH': '/usr/bin:/bin:/usr/sbin:/sbin', 'PROMPT_COMMAND': 'printf "\\033]0;\\007"', 'PS1': ' \\s-\\v\\$ ', 'LANG': 'C.UTF-8', 'SYSTEMD_NSPAWN_TMPFS_TMP': '0', 'SYSTEMD_SECCOMP': '0'} and shell False Building target platforms: x86_64 Building for target x86_64 setting SOURCE_DATE_EPOCH=1776384000 Executing(%generate_buildrequires): /bin/sh -e /var/tmp/rpm-tmp.oMh1et + umask 022 + cd /builddir/build/BUILD + cd confluent_kafka-2.12.0 + echo pyproject-rpm-macros + echo python3-devel + echo 'python3dist(packaging)' + echo 'python3dist(pip) >= 19' + '[' -f pyproject.toml ']' + echo '(python3dist(tomli) if python3-devel < 3.11)' + rm -rfv '*.dist-info/' + '[' -f /usr/bin/python3 ']' + mkdir -p /builddir/build/BUILD/confluent_kafka-2.12.0/.pyproject-builddir + echo -n + CFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -march=x86-64-v2 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection' + CXXFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -march=x86-64-v2 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection' + FFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -march=x86-64-v2 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -I/usr/lib64/gfortran/modules' + FCFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -march=x86-64-v2 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -I/usr/lib64/gfortran/modules' + LDFLAGS='-Wl,-z,relro -Wl,--as-needed -Wl,-z,now -specs=/usr/lib/rpm/redhat/redhat-hardened-ld -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 ' + LT_SYS_LIBRARY_PATH=/usr/lib64: + CC=gcc + CXX=g++ + TMPDIR=/builddir/build/BUILD/confluent_kafka-2.12.0/.pyproject-builddir + RPM_TOXENV=py39 + HOSTNAME=rpmbuild + /usr/bin/python3 -Bs /usr/lib/rpm/redhat/pyproject_buildrequires.py --generate-extras --python3_pkgversion 3 --wheeldir /builddir/build/BUILD/confluent_kafka-2.12.0/pyproject-wheeldir --output /builddir/build/BUILD/python-confluent-kafka-2.12.0-1.el9.x86_64-pyproject-buildrequires Handling setuptools>=62 from build-system.requires Requirement satisfied: setuptools>=62 (installed: setuptools 65.5.1) Handling wheel from build-system.requires Requirement satisfied: wheel (installed: wheel 0.36.2) /usr/lib/python3.9/site-packages/setuptools/config/pyprojecttoml.py:108: _BetaConfiguration: Support for `[tool.setuptools]` in `pyproject.toml` is still *beta*. warnings.warn(msg, _BetaConfiguration) running egg_info writing src/confluent_kafka.egg-info/PKG-INFO writing dependency_links to src/confluent_kafka.egg-info/dependency_links.txt writing requirements to src/confluent_kafka.egg-info/requires.txt writing top-level names to src/confluent_kafka.egg-info/top_level.txt reading manifest file 'src/confluent_kafka.egg-info/SOURCES.txt' reading manifest template 'MANIFEST.in' no previously-included directories found matching 'tests' no previously-included directories found matching 'docs' adding license file 'LICENSE' writing manifest file 'src/confluent_kafka.egg-info/SOURCES.txt' Handling wheel from get_requires_for_build_wheel Requirement satisfied: wheel (installed: wheel 0.36.2) /usr/lib/python3.9/site-packages/setuptools/config/pyprojecttoml.py:108: _BetaConfiguration: Support for `[tool.setuptools]` in `pyproject.toml` is still *beta*. warnings.warn(msg, _BetaConfiguration) running dist_info writing confluent_kafka.egg-info/PKG-INFO writing dependency_links to confluent_kafka.egg-info/dependency_links.txt writing requirements to confluent_kafka.egg-info/requires.txt writing top-level names to confluent_kafka.egg-info/top_level.txt reading manifest file 'confluent_kafka.egg-info/SOURCES.txt' reading manifest template 'MANIFEST.in' no previously-included directories found matching 'tests' no previously-included directories found matching 'docs' adding license file 'LICENSE' writing manifest file 'confluent_kafka.egg-info/SOURCES.txt' creating '/builddir/build/BUILD/confluent_kafka-2.12.0/confluent_kafka-2.12.0.dist-info' adding license file "LICENSE" (matched pattern "LICEN[CS]E*") Handling tomli ; python_version < "3.13" from hook generated metadata: Requires-Dist (confluent-kafka) Requirement satisfied: tomli ; python_version < "3.13" (installed: tomli 2.0.1) Handling psutil ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: psutil ; extra == 'all' Handling opentelemetry-distro ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: opentelemetry-distro ; extra == 'all' Handling opentelemetry-exporter-otlp ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: opentelemetry-exporter-otlp ; extra == 'all' Handling sphinx ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: sphinx ; extra == 'all' Handling sphinx-rtd-theme ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: sphinx-rtd-theme ; extra == 'all' Handling pandoc ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: pandoc ; extra == 'all' Handling confluent-kafka ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: confluent-kafka ; extra == 'all' Handling fastapi ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: fastapi ; extra == 'all' Handling pydantic ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: pydantic ; extra == 'all' Handling uvicorn ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: uvicorn ; extra == 'all' Handling six ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: six ; extra == 'all' Handling attrs ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: attrs ; extra == 'all' Handling cachetools ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: cachetools ; extra == 'all' Handling httpx (>=0.26) ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: httpx (>=0.26) ; extra == 'all' Handling authlib (>=1.0.0) ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: authlib (>=1.0.0) ; extra == 'all' Handling requests ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: requests ; extra == 'all' Handling avro (<2,>=1.11.1) ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: avro (<2,>=1.11.1) ; extra == 'all' Handling pyrsistent ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: pyrsistent ; extra == 'all' Handling jsonschema ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: jsonschema ; extra == 'all' Handling orjson (>=3.10) ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: orjson (>=3.10) ; extra == 'all' Handling googleapis-common-protos ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: googleapis-common-protos ; extra == 'all' Handling protobuf ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: protobuf ; extra == 'all' Handling azure-identity ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: azure-identity ; extra == 'all' Handling azure-keyvault-keys ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: azure-keyvault-keys ; extra == 'all' Handling boto3 ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: boto3 ; extra == 'all' Handling cel-python (>=0.4.0) ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: cel-python (>=0.4.0) ; extra == 'all' Handling google-auth ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: google-auth ; extra == 'all' Handling google-api-core ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: google-api-core ; extra == 'all' Handling google-cloud-kms ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: google-cloud-kms ; extra == 'all' Handling hkdf (==0.0.3) ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: hkdf (==0.0.3) ; extra == 'all' Handling hvac ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: hvac ; extra == 'all' Handling jsonata-python ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: jsonata-python ; extra == 'all' Handling pyyaml (>=6.0.0) ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: pyyaml (>=6.0.0) ; extra == 'all' Handling tink ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: tink ; extra == 'all' Handling urllib3 (<3) ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: urllib3 (<3) ; extra == 'all' Handling flake8 ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: flake8 ; extra == 'all' Handling orjson ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: orjson ; extra == 'all' Handling pytest ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: pytest ; extra == 'all' Handling pytest-timeout ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: pytest-timeout ; extra == 'all' Handling requests-mock ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: requests-mock ; extra == 'all' Handling respx ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: respx ; extra == 'all' Handling pytest-cov ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: pytest-cov ; extra == 'all' Handling pluggy (<1.6.0) ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: pluggy (<1.6.0) ; extra == 'all' Handling pytest-asyncio ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: pytest-asyncio ; extra == 'all' Handling async-timeout ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: async-timeout ; extra == 'all' Handling attrs (>=21.2.0) ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: attrs (>=21.2.0) ; extra == 'all' Handling cachetools (>=5.5.0) ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: cachetools (>=5.5.0) ; extra == 'all' Handling certifi ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: certifi ; extra == 'all' Handling boto3 (>=1.35) ; extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: boto3 (>=1.35) ; extra == 'all' Handling tomli ; (python_version < "3.13") and extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: tomli ; (python_version < "3.13") and extra == 'all' Handling fastavro (<1.8.0) ; (python_version == "3.7") and extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: fastavro (<1.8.0) ; (python_version == "3.7") and extra == 'all' Handling fastavro (<2) ; (python_version > "3.7") and extra == 'all' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: fastavro (<2) ; (python_version > "3.7") and extra == 'all' Handling requests ; extra == 'avro' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: requests ; extra == 'avro' Handling avro (<2,>=1.11.1) ; extra == 'avro' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: avro (<2,>=1.11.1) ; extra == 'avro' Handling attrs (>=21.2.0) ; extra == 'avro' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: attrs (>=21.2.0) ; extra == 'avro' Handling cachetools (>=5.5.0) ; extra == 'avro' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: cachetools (>=5.5.0) ; extra == 'avro' Handling certifi ; extra == 'avro' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: certifi ; extra == 'avro' Handling httpx (>=0.26) ; extra == 'avro' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: httpx (>=0.26) ; extra == 'avro' Handling authlib (>=1.0.0) ; extra == 'avro' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: authlib (>=1.0.0) ; extra == 'avro' Handling fastavro (<1.8.0) ; (python_version == "3.7") and extra == 'avro' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: fastavro (<1.8.0) ; (python_version == "3.7") and extra == 'avro' Handling fastavro (<2) ; (python_version > "3.7") and extra == 'avro' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: fastavro (<2) ; (python_version > "3.7") and extra == 'avro' Handling sphinx ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: sphinx ; extra == 'dev' Handling sphinx-rtd-theme ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: sphinx-rtd-theme ; extra == 'dev' Handling pandoc ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: pandoc ; extra == 'dev' Handling confluent-kafka ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: confluent-kafka ; extra == 'dev' Handling fastapi ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: fastapi ; extra == 'dev' Handling pydantic ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: pydantic ; extra == 'dev' Handling uvicorn ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: uvicorn ; extra == 'dev' Handling six ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: six ; extra == 'dev' Handling attrs ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: attrs ; extra == 'dev' Handling cachetools ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: cachetools ; extra == 'dev' Handling httpx (>=0.26) ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: httpx (>=0.26) ; extra == 'dev' Handling authlib (>=1.0.0) ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: authlib (>=1.0.0) ; extra == 'dev' Handling requests ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: requests ; extra == 'dev' Handling avro (<2,>=1.11.1) ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: avro (<2,>=1.11.1) ; extra == 'dev' Handling pyrsistent ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: pyrsistent ; extra == 'dev' Handling jsonschema ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: jsonschema ; extra == 'dev' Handling orjson (>=3.10) ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: orjson (>=3.10) ; extra == 'dev' Handling googleapis-common-protos ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: googleapis-common-protos ; extra == 'dev' Handling protobuf ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: protobuf ; extra == 'dev' Handling azure-identity ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: azure-identity ; extra == 'dev' Handling azure-keyvault-keys ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: azure-keyvault-keys ; extra == 'dev' Handling boto3 ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: boto3 ; extra == 'dev' Handling cel-python (>=0.4.0) ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: cel-python (>=0.4.0) ; extra == 'dev' Handling google-auth ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: google-auth ; extra == 'dev' Handling google-api-core ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: google-api-core ; extra == 'dev' Handling google-cloud-kms ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: google-cloud-kms ; extra == 'dev' Handling hkdf (==0.0.3) ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: hkdf (==0.0.3) ; extra == 'dev' Handling hvac ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: hvac ; extra == 'dev' Handling jsonata-python ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: jsonata-python ; extra == 'dev' Handling pyyaml (>=6.0.0) ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: pyyaml (>=6.0.0) ; extra == 'dev' Handling tink ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: tink ; extra == 'dev' Handling urllib3 (<3) ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: urllib3 (<3) ; extra == 'dev' Handling flake8 ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: flake8 ; extra == 'dev' Handling orjson ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: orjson ; extra == 'dev' Handling pytest ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: pytest ; extra == 'dev' Handling pytest-timeout ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: pytest-timeout ; extra == 'dev' Handling requests-mock ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: requests-mock ; extra == 'dev' Handling respx ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: respx ; extra == 'dev' Handling pytest-cov ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: pytest-cov ; extra == 'dev' Handling pluggy (<1.6.0) ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: pluggy (<1.6.0) ; extra == 'dev' Handling pytest-asyncio ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: pytest-asyncio ; extra == 'dev' Handling async-timeout ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: async-timeout ; extra == 'dev' Handling attrs (>=21.2.0) ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: attrs (>=21.2.0) ; extra == 'dev' Handling cachetools (>=5.5.0) ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: cachetools (>=5.5.0) ; extra == 'dev' Handling certifi ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: certifi ; extra == 'dev' Handling boto3 (>=1.35) ; extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: boto3 (>=1.35) ; extra == 'dev' Handling tomli ; (python_version < "3.13") and extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: tomli ; (python_version < "3.13") and extra == 'dev' Handling fastavro (<1.8.0) ; (python_version == "3.7") and extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: fastavro (<1.8.0) ; (python_version == "3.7") and extra == 'dev' Handling fastavro (<2) ; (python_version > "3.7") and extra == 'dev' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: fastavro (<2) ; (python_version > "3.7") and extra == 'dev' Handling sphinx ; extra == 'docs' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: sphinx ; extra == 'docs' Handling sphinx-rtd-theme ; extra == 'docs' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: sphinx-rtd-theme ; extra == 'docs' Handling pandoc ; extra == 'docs' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: pandoc ; extra == 'docs' Handling attrs (>=21.2.0) ; extra == 'docs' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: attrs (>=21.2.0) ; extra == 'docs' Handling cachetools (>=5.5.0) ; extra == 'docs' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: cachetools (>=5.5.0) ; extra == 'docs' Handling certifi ; extra == 'docs' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: certifi ; extra == 'docs' Handling httpx (>=0.26) ; extra == 'docs' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: httpx (>=0.26) ; extra == 'docs' Handling authlib (>=1.0.0) ; extra == 'docs' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: authlib (>=1.0.0) ; extra == 'docs' Handling azure-identity ; extra == 'docs' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: azure-identity ; extra == 'docs' Handling azure-keyvault-keys ; extra == 'docs' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: azure-keyvault-keys ; extra == 'docs' Handling boto3 (>=1.35) ; extra == 'docs' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: boto3 (>=1.35) ; extra == 'docs' Handling cel-python (>=0.4.0) ; extra == 'docs' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: cel-python (>=0.4.0) ; extra == 'docs' Handling google-auth ; extra == 'docs' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: google-auth ; extra == 'docs' Handling google-api-core ; extra == 'docs' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: google-api-core ; extra == 'docs' Handling google-cloud-kms ; extra == 'docs' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: google-cloud-kms ; extra == 'docs' Handling hkdf (==0.0.3) ; extra == 'docs' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: hkdf (==0.0.3) ; extra == 'docs' Handling hvac ; extra == 'docs' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: hvac ; extra == 'docs' Handling jsonata-python ; extra == 'docs' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: jsonata-python ; extra == 'docs' Handling pyyaml (>=6.0.0) ; extra == 'docs' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: pyyaml (>=6.0.0) ; extra == 'docs' Handling tink ; extra == 'docs' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: tink ; extra == 'docs' Handling requests ; extra == 'docs' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: requests ; extra == 'docs' Handling avro (<2,>=1.11.1) ; extra == 'docs' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: avro (<2,>=1.11.1) ; extra == 'docs' Handling pyrsistent ; extra == 'docs' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: pyrsistent ; extra == 'docs' Handling jsonschema ; extra == 'docs' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: jsonschema ; extra == 'docs' Handling orjson (>=3.10) ; extra == 'docs' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: orjson (>=3.10) ; extra == 'docs' Handling googleapis-common-protos ; extra == 'docs' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: googleapis-common-protos ; extra == 'docs' Handling protobuf ; extra == 'docs' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: protobuf ; extra == 'docs' Handling tomli ; (python_version < "3.13") and extra == 'docs' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: tomli ; (python_version < "3.13") and extra == 'docs' Handling fastavro (<1.8.0) ; (python_version == "3.7") and extra == 'docs' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: fastavro (<1.8.0) ; (python_version == "3.7") and extra == 'docs' Handling fastavro (<2) ; (python_version > "3.7") and extra == 'docs' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: fastavro (<2) ; (python_version > "3.7") and extra == 'docs' Handling confluent-kafka ; extra == 'examples' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: confluent-kafka ; extra == 'examples' Handling fastapi ; extra == 'examples' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: fastapi ; extra == 'examples' Handling pydantic ; extra == 'examples' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: pydantic ; extra == 'examples' Handling uvicorn ; extra == 'examples' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: uvicorn ; extra == 'examples' Handling six ; extra == 'examples' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: six ; extra == 'examples' Handling attrs ; extra == 'examples' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: attrs ; extra == 'examples' Handling cachetools ; extra == 'examples' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: cachetools ; extra == 'examples' Handling httpx (>=0.26) ; extra == 'examples' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: httpx (>=0.26) ; extra == 'examples' Handling authlib (>=1.0.0) ; extra == 'examples' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: authlib (>=1.0.0) ; extra == 'examples' Handling requests ; extra == 'examples' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: requests ; extra == 'examples' Handling avro (<2,>=1.11.1) ; extra == 'examples' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: avro (<2,>=1.11.1) ; extra == 'examples' Handling pyrsistent ; extra == 'examples' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: pyrsistent ; extra == 'examples' Handling jsonschema ; extra == 'examples' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: jsonschema ; extra == 'examples' Handling orjson (>=3.10) ; extra == 'examples' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: orjson (>=3.10) ; extra == 'examples' Handling googleapis-common-protos ; extra == 'examples' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: googleapis-common-protos ; extra == 'examples' Handling protobuf ; extra == 'examples' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: protobuf ; extra == 'examples' Handling azure-identity ; extra == 'examples' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: azure-identity ; extra == 'examples' Handling azure-keyvault-keys ; extra == 'examples' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: azure-keyvault-keys ; extra == 'examples' Handling boto3 ; extra == 'examples' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: boto3 ; extra == 'examples' Handling cel-python (>=0.4.0) ; extra == 'examples' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: cel-python (>=0.4.0) ; extra == 'examples' Handling google-auth ; extra == 'examples' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: google-auth ; extra == 'examples' Handling google-api-core ; extra == 'examples' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: google-api-core ; extra == 'examples' Handling google-cloud-kms ; extra == 'examples' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: google-cloud-kms ; extra == 'examples' Handling hkdf (==0.0.3) ; extra == 'examples' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: hkdf (==0.0.3) ; extra == 'examples' Handling hvac ; extra == 'examples' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: hvac ; extra == 'examples' Handling jsonata-python ; extra == 'examples' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: jsonata-python ; extra == 'examples' Handling pyyaml (>=6.0.0) ; extra == 'examples' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: pyyaml (>=6.0.0) ; extra == 'examples' Handling tink ; extra == 'examples' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: tink ; extra == 'examples' Handling fastavro (<1.8.0) ; (python_version == "3.7") and extra == 'examples' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: fastavro (<1.8.0) ; (python_version == "3.7") and extra == 'examples' Handling fastavro (<2) ; (python_version > "3.7") and extra == 'examples' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: fastavro (<2) ; (python_version > "3.7") and extra == 'examples' Handling pyrsistent ; extra == 'json' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: pyrsistent ; extra == 'json' Handling jsonschema ; extra == 'json' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: jsonschema ; extra == 'json' Handling orjson (>=3.10) ; extra == 'json' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: orjson (>=3.10) ; extra == 'json' Handling attrs (>=21.2.0) ; extra == 'json' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: attrs (>=21.2.0) ; extra == 'json' Handling cachetools (>=5.5.0) ; extra == 'json' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: cachetools (>=5.5.0) ; extra == 'json' Handling certifi ; extra == 'json' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: certifi ; extra == 'json' Handling httpx (>=0.26) ; extra == 'json' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: httpx (>=0.26) ; extra == 'json' Handling authlib (>=1.0.0) ; extra == 'json' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: authlib (>=1.0.0) ; extra == 'json' Handling googleapis-common-protos ; extra == 'protobuf' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: googleapis-common-protos ; extra == 'protobuf' Handling protobuf ; extra == 'protobuf' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: protobuf ; extra == 'protobuf' Handling attrs (>=21.2.0) ; extra == 'protobuf' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: attrs (>=21.2.0) ; extra == 'protobuf' Handling cachetools (>=5.5.0) ; extra == 'protobuf' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: cachetools (>=5.5.0) ; extra == 'protobuf' Handling certifi ; extra == 'protobuf' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: certifi ; extra == 'protobuf' Handling httpx (>=0.26) ; extra == 'protobuf' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: httpx (>=0.26) ; extra == 'protobuf' Handling authlib (>=1.0.0) ; extra == 'protobuf' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: authlib (>=1.0.0) ; extra == 'protobuf' Handling azure-identity ; extra == 'rules' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: azure-identity ; extra == 'rules' Handling azure-keyvault-keys ; extra == 'rules' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: azure-keyvault-keys ; extra == 'rules' Handling boto3 (>=1.35) ; extra == 'rules' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: boto3 (>=1.35) ; extra == 'rules' Handling cel-python (>=0.4.0) ; extra == 'rules' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: cel-python (>=0.4.0) ; extra == 'rules' Handling google-auth ; extra == 'rules' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: google-auth ; extra == 'rules' Handling google-api-core ; extra == 'rules' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: google-api-core ; extra == 'rules' Handling google-cloud-kms ; extra == 'rules' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: google-cloud-kms ; extra == 'rules' Handling hkdf (==0.0.3) ; extra == 'rules' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: hkdf (==0.0.3) ; extra == 'rules' Handling hvac ; extra == 'rules' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: hvac ; extra == 'rules' Handling jsonata-python ; extra == 'rules' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: jsonata-python ; extra == 'rules' Handling pyyaml (>=6.0.0) ; extra == 'rules' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: pyyaml (>=6.0.0) ; extra == 'rules' Handling tink ; extra == 'rules' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: tink ; extra == 'rules' Handling attrs (>=21.2.0) ; extra == 'rules' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: attrs (>=21.2.0) ; extra == 'rules' Handling cachetools (>=5.5.0) ; extra == 'rules' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: cachetools (>=5.5.0) ; extra == 'rules' Handling certifi ; extra == 'rules' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: certifi ; extra == 'rules' Handling httpx (>=0.26) ; extra == 'rules' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: httpx (>=0.26) ; extra == 'rules' Handling authlib (>=1.0.0) ; extra == 'rules' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: authlib (>=1.0.0) ; extra == 'rules' Handling attrs (>=21.2.0) ; extra == 'schema_registry' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: attrs (>=21.2.0) ; extra == 'schema_registry' Handling cachetools (>=5.5.0) ; extra == 'schema_registry' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: cachetools (>=5.5.0) ; extra == 'schema_registry' Handling certifi ; extra == 'schema_registry' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: certifi ; extra == 'schema_registry' Handling httpx (>=0.26) ; extra == 'schema_registry' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: httpx (>=0.26) ; extra == 'schema_registry' Handling authlib (>=1.0.0) ; extra == 'schema_registry' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: authlib (>=1.0.0) ; extra == 'schema_registry' Handling attrs (>=21.2.0) ; extra == 'schemaregistry' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: attrs (>=21.2.0) ; extra == 'schemaregistry' Handling cachetools (>=5.5.0) ; extra == 'schemaregistry' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: cachetools (>=5.5.0) ; extra == 'schemaregistry' Handling certifi ; extra == 'schemaregistry' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: certifi ; extra == 'schemaregistry' Handling httpx (>=0.26) ; extra == 'schemaregistry' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: httpx (>=0.26) ; extra == 'schemaregistry' Handling authlib (>=1.0.0) ; extra == 'schemaregistry' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: authlib (>=1.0.0) ; extra == 'schemaregistry' Handling psutil ; extra == 'soaktest' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: psutil ; extra == 'soaktest' Handling opentelemetry-distro ; extra == 'soaktest' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: opentelemetry-distro ; extra == 'soaktest' Handling opentelemetry-exporter-otlp ; extra == 'soaktest' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: opentelemetry-exporter-otlp ; extra == 'soaktest' Handling urllib3 (<3) ; extra == 'tests' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: urllib3 (<3) ; extra == 'tests' Handling flake8 ; extra == 'tests' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: flake8 ; extra == 'tests' Handling orjson ; extra == 'tests' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: orjson ; extra == 'tests' Handling pytest ; extra == 'tests' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: pytest ; extra == 'tests' Handling pytest-timeout ; extra == 'tests' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: pytest-timeout ; extra == 'tests' Handling requests-mock ; extra == 'tests' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: requests-mock ; extra == 'tests' Handling respx ; extra == 'tests' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: respx ; extra == 'tests' Handling pytest-cov ; extra == 'tests' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: pytest-cov ; extra == 'tests' Handling pluggy (<1.6.0) ; extra == 'tests' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: pluggy (<1.6.0) ; extra == 'tests' Handling pytest-asyncio ; extra == 'tests' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: pytest-asyncio ; extra == 'tests' Handling async-timeout ; extra == 'tests' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: async-timeout ; extra == 'tests' Handling attrs (>=21.2.0) ; extra == 'tests' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: attrs (>=21.2.0) ; extra == 'tests' Handling cachetools (>=5.5.0) ; extra == 'tests' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: cachetools (>=5.5.0) ; extra == 'tests' Handling certifi ; extra == 'tests' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: certifi ; extra == 'tests' Handling httpx (>=0.26) ; extra == 'tests' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: httpx (>=0.26) ; extra == 'tests' Handling authlib (>=1.0.0) ; extra == 'tests' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: authlib (>=1.0.0) ; extra == 'tests' Handling azure-identity ; extra == 'tests' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: azure-identity ; extra == 'tests' Handling azure-keyvault-keys ; extra == 'tests' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: azure-keyvault-keys ; extra == 'tests' Handling boto3 (>=1.35) ; extra == 'tests' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: boto3 (>=1.35) ; extra == 'tests' Handling cel-python (>=0.4.0) ; extra == 'tests' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: cel-python (>=0.4.0) ; extra == 'tests' Handling google-auth ; extra == 'tests' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: google-auth ; extra == 'tests' Handling google-api-core ; extra == 'tests' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: google-api-core ; extra == 'tests' Handling google-cloud-kms ; extra == 'tests' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: google-cloud-kms ; extra == 'tests' Handling hkdf (==0.0.3) ; extra == 'tests' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: hkdf (==0.0.3) ; extra == 'tests' Handling hvac ; extra == 'tests' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: hvac ; extra == 'tests' Handling jsonata-python ; extra == 'tests' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: jsonata-python ; extra == 'tests' Handling pyyaml (>=6.0.0) ; extra == 'tests' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: pyyaml (>=6.0.0) ; extra == 'tests' Handling tink ; extra == 'tests' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: tink ; extra == 'tests' Handling requests ; extra == 'tests' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: requests ; extra == 'tests' Handling avro (<2,>=1.11.1) ; extra == 'tests' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: avro (<2,>=1.11.1) ; extra == 'tests' Handling pyrsistent ; extra == 'tests' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: pyrsistent ; extra == 'tests' Handling jsonschema ; extra == 'tests' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: jsonschema ; extra == 'tests' Handling orjson (>=3.10) ; extra == 'tests' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: orjson (>=3.10) ; extra == 'tests' Handling googleapis-common-protos ; extra == 'tests' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: googleapis-common-protos ; extra == 'tests' Handling protobuf ; extra == 'tests' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: protobuf ; extra == 'tests' Handling fastavro (<1.8.0) ; (python_version == "3.7") and extra == 'tests' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: fastavro (<1.8.0) ; (python_version == "3.7") and extra == 'tests' Handling fastavro (<2) ; (python_version > "3.7") and extra == 'tests' from hook generated metadata: Requires-Dist (confluent-kafka) Ignoring alien requirement: fastavro (<2) ; (python_version > "3.7") and extra == 'tests' + cat /builddir/build/BUILD/python-confluent-kafka-2.12.0-1.el9.x86_64-pyproject-buildrequires + rm -rfv confluent_kafka-2.12.0.dist-info/ removed 'confluent_kafka-2.12.0.dist-info/top_level.txt' removed 'confluent_kafka-2.12.0.dist-info/METADATA' removed 'confluent_kafka-2.12.0.dist-info/LICENSE' removed directory 'confluent_kafka-2.12.0.dist-info/' + RPM_EC=0 ++ jobs -p + exit 0 Executing(%build): /bin/sh -e /var/tmp/rpm-tmp.JUQQCB + umask 022 + cd /builddir/build/BUILD + cd confluent_kafka-2.12.0 + mkdir -p /builddir/build/BUILD/confluent_kafka-2.12.0/.pyproject-builddir + CFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -march=x86-64-v2 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection' + CXXFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -march=x86-64-v2 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection' + FFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -march=x86-64-v2 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -I/usr/lib64/gfortran/modules' + FCFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -march=x86-64-v2 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -I/usr/lib64/gfortran/modules' + LDFLAGS='-Wl,-z,relro -Wl,--as-needed -Wl,-z,now -specs=/usr/lib/rpm/redhat/redhat-hardened-ld -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 ' + LT_SYS_LIBRARY_PATH=/usr/lib64: + CC=gcc + CXX=g++ + TMPDIR=/builddir/build/BUILD/confluent_kafka-2.12.0/.pyproject-builddir + /usr/bin/python3 -Bs /usr/lib/rpm/redhat/pyproject_wheel.py /builddir/build/BUILD/confluent_kafka-2.12.0/pyproject-wheeldir Processing /builddir/build/BUILD/confluent_kafka-2.12.0 Preparing metadata (pyproject.toml): started Running command /usr/bin/python3 /usr/lib/python3.9/site-packages/pip/_vendor/pep517/in_process/_in_process.py prepare_metadata_for_build_wheel /builddir/build/BUILD/confluent_kafka-2.12.0/.pyproject-builddir/tmpb6tis3xv /usr/lib/python3.9/site-packages/setuptools/config/pyprojecttoml.py:108: _BetaConfiguration: Support for `[tool.setuptools]` in `pyproject.toml` is still *beta*. warnings.warn(msg, _BetaConfiguration) running dist_info creating /builddir/build/BUILD/confluent_kafka-2.12.0/.pyproject-builddir/pip-modern-metadata-b7xo_xwa/confluent_kafka.egg-info writing /builddir/build/BUILD/confluent_kafka-2.12.0/.pyproject-builddir/pip-modern-metadata-b7xo_xwa/confluent_kafka.egg-info/PKG-INFO writing dependency_links to /builddir/build/BUILD/confluent_kafka-2.12.0/.pyproject-builddir/pip-modern-metadata-b7xo_xwa/confluent_kafka.egg-info/dependency_links.txt writing requirements to /builddir/build/BUILD/confluent_kafka-2.12.0/.pyproject-builddir/pip-modern-metadata-b7xo_xwa/confluent_kafka.egg-info/requires.txt writing top-level names to /builddir/build/BUILD/confluent_kafka-2.12.0/.pyproject-builddir/pip-modern-metadata-b7xo_xwa/confluent_kafka.egg-info/top_level.txt writing manifest file '/builddir/build/BUILD/confluent_kafka-2.12.0/.pyproject-builddir/pip-modern-metadata-b7xo_xwa/confluent_kafka.egg-info/SOURCES.txt' reading manifest file '/builddir/build/BUILD/confluent_kafka-2.12.0/.pyproject-builddir/pip-modern-metadata-b7xo_xwa/confluent_kafka.egg-info/SOURCES.txt' reading manifest template 'MANIFEST.in' no previously-included directories found matching 'tests' no previously-included directories found matching 'docs' adding license file 'LICENSE' writing manifest file '/builddir/build/BUILD/confluent_kafka-2.12.0/.pyproject-builddir/pip-modern-metadata-b7xo_xwa/confluent_kafka.egg-info/SOURCES.txt' creating '/builddir/build/BUILD/confluent_kafka-2.12.0/.pyproject-builddir/pip-modern-metadata-b7xo_xwa/confluent_kafka-2.12.0.dist-info' adding license file "LICENSE" (matched pattern "LICEN[CS]E*") Preparing metadata (pyproject.toml): finished with status 'done' Building wheels for collected packages: confluent-kafka Building wheel for confluent-kafka (pyproject.toml): started Running command /usr/bin/python3 /usr/lib/python3.9/site-packages/pip/_vendor/pep517/in_process/_in_process.py build_wheel /builddir/build/BUILD/confluent_kafka-2.12.0/.pyproject-builddir/tmpq13em2tb /usr/lib/python3.9/site-packages/setuptools/config/pyprojecttoml.py:108: _BetaConfiguration: Support for `[tool.setuptools]` in `pyproject.toml` is still *beta*. warnings.warn(msg, _BetaConfiguration) running bdist_wheel running build running build_py creating build creating build/lib.linux-x86_64-cpython-39 creating build/lib.linux-x86_64-cpython-39/confluent_kafka copying src/confluent_kafka/serializing_producer.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka copying src/confluent_kafka/error.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka copying src/confluent_kafka/__init__.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka copying src/confluent_kafka/deserializing_consumer.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka creating build/lib.linux-x86_64-cpython-39/confluent_kafka/_util copying src/confluent_kafka/_util/validation_util.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/_util copying src/confluent_kafka/_util/conversion_util.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/_util copying src/confluent_kafka/_util/__init__.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/_util creating build/lib.linux-x86_64-cpython-39/confluent_kafka/serialization copying src/confluent_kafka/serialization/__init__.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/serialization creating build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry copying src/confluent_kafka/schema_registry/rule_registry.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry copying src/confluent_kafka/schema_registry/schema_registry_client.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry copying src/confluent_kafka/schema_registry/error.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry copying src/confluent_kafka/schema_registry/json_schema.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry copying src/confluent_kafka/schema_registry/serde.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry copying src/confluent_kafka/schema_registry/avro.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry copying src/confluent_kafka/schema_registry/protobuf.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry copying src/confluent_kafka/schema_registry/__init__.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry copying src/confluent_kafka/schema_registry/wildcard_matcher.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry creating build/lib.linux-x86_64-cpython-39/confluent_kafka/_model copying src/confluent_kafka/_model/__init__.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/_model creating build/lib.linux-x86_64-cpython-39/confluent_kafka/experimental copying src/confluent_kafka/experimental/__init__.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/experimental creating build/lib.linux-x86_64-cpython-39/confluent_kafka/admin copying src/confluent_kafka/admin/_scram.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/admin copying src/confluent_kafka/admin/_acl.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/admin copying src/confluent_kafka/admin/_listoffsets.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/admin copying src/confluent_kafka/admin/_resource.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/admin copying src/confluent_kafka/admin/_config.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/admin copying src/confluent_kafka/admin/_cluster.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/admin copying src/confluent_kafka/admin/_group.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/admin copying src/confluent_kafka/admin/_records.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/admin copying src/confluent_kafka/admin/_topic.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/admin copying src/confluent_kafka/admin/__init__.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/admin copying src/confluent_kafka/admin/_metadata.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/admin creating build/lib.linux-x86_64-cpython-39/confluent_kafka/kafkatest copying src/confluent_kafka/kafkatest/verifiable_client.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/kafkatest copying src/confluent_kafka/kafkatest/verifiable_consumer.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/kafkatest copying src/confluent_kafka/kafkatest/verifiable_producer.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/kafkatest copying src/confluent_kafka/kafkatest/__init__.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/kafkatest creating build/lib.linux-x86_64-cpython-39/confluent_kafka/avro copying src/confluent_kafka/avro/error.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/avro copying src/confluent_kafka/avro/load.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/avro copying src/confluent_kafka/avro/cached_schema_registry_client.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/avro copying src/confluent_kafka/avro/__init__.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/avro creating build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/_sync copying src/confluent_kafka/schema_registry/_sync/mock_schema_registry_client.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/_sync copying src/confluent_kafka/schema_registry/_sync/schema_registry_client.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/_sync copying src/confluent_kafka/schema_registry/_sync/json_schema.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/_sync copying src/confluent_kafka/schema_registry/_sync/serde.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/_sync copying src/confluent_kafka/schema_registry/_sync/avro.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/_sync copying src/confluent_kafka/schema_registry/_sync/protobuf.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/_sync copying src/confluent_kafka/schema_registry/_sync/__init__.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/_sync creating build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/_async copying src/confluent_kafka/schema_registry/_async/mock_schema_registry_client.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/_async copying src/confluent_kafka/schema_registry/_async/schema_registry_client.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/_async copying src/confluent_kafka/schema_registry/_async/json_schema.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/_async copying src/confluent_kafka/schema_registry/_async/serde.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/_async copying src/confluent_kafka/schema_registry/_async/avro.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/_async copying src/confluent_kafka/schema_registry/_async/protobuf.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/_async copying src/confluent_kafka/schema_registry/_async/__init__.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/_async creating build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/rules copying src/confluent_kafka/schema_registry/rules/__init__.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/rules creating build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/common copying src/confluent_kafka/schema_registry/common/schema_registry_client.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/common copying src/confluent_kafka/schema_registry/common/json_schema.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/common copying src/confluent_kafka/schema_registry/common/serde.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/common copying src/confluent_kafka/schema_registry/common/avro.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/common copying src/confluent_kafka/schema_registry/common/protobuf.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/common copying src/confluent_kafka/schema_registry/common/__init__.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/common creating build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/confluent copying src/confluent_kafka/schema_registry/confluent/meta_pb2.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/confluent copying src/confluent_kafka/schema_registry/confluent/__init__.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/confluent creating build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/rules/jsonata copying src/confluent_kafka/schema_registry/rules/jsonata/jsonata_executor.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/rules/jsonata copying src/confluent_kafka/schema_registry/rules/jsonata/__init__.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/rules/jsonata creating build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/rules/encryption copying src/confluent_kafka/schema_registry/rules/encryption/encrypt_executor.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/rules/encryption copying src/confluent_kafka/schema_registry/rules/encryption/__init__.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/rules/encryption copying src/confluent_kafka/schema_registry/rules/encryption/kms_driver_registry.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/rules/encryption creating build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/rules/cel copying src/confluent_kafka/schema_registry/rules/cel/extra_func.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/rules/cel copying src/confluent_kafka/schema_registry/rules/cel/string_format.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/rules/cel copying src/confluent_kafka/schema_registry/rules/cel/cel_executor.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/rules/cel copying src/confluent_kafka/schema_registry/rules/cel/cel_field_presence.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/rules/cel copying src/confluent_kafka/schema_registry/rules/cel/constraints.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/rules/cel copying src/confluent_kafka/schema_registry/rules/cel/__init__.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/rules/cel copying src/confluent_kafka/schema_registry/rules/cel/cel_field_executor.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/rules/cel creating build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/rules/encryption/gcpkms copying src/confluent_kafka/schema_registry/rules/encryption/gcpkms/gcp_driver.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/rules/encryption/gcpkms copying src/confluent_kafka/schema_registry/rules/encryption/gcpkms/gcp_client.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/rules/encryption/gcpkms copying src/confluent_kafka/schema_registry/rules/encryption/gcpkms/__init__.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/rules/encryption/gcpkms creating build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/rules/encryption/hcvault copying src/confluent_kafka/schema_registry/rules/encryption/hcvault/hcvault_client.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/rules/encryption/hcvault copying src/confluent_kafka/schema_registry/rules/encryption/hcvault/hcvault_driver.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/rules/encryption/hcvault copying src/confluent_kafka/schema_registry/rules/encryption/hcvault/__init__.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/rules/encryption/hcvault creating build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/rules/encryption/awskms copying src/confluent_kafka/schema_registry/rules/encryption/awskms/aws_driver.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/rules/encryption/awskms copying src/confluent_kafka/schema_registry/rules/encryption/awskms/__init__.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/rules/encryption/awskms creating build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/rules/encryption/localkms copying src/confluent_kafka/schema_registry/rules/encryption/localkms/local_driver.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/rules/encryption/localkms copying src/confluent_kafka/schema_registry/rules/encryption/localkms/local_client.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/rules/encryption/localkms copying src/confluent_kafka/schema_registry/rules/encryption/localkms/__init__.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/rules/encryption/localkms creating build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/rules/encryption/azurekms copying src/confluent_kafka/schema_registry/rules/encryption/azurekms/azure_aead.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/rules/encryption/azurekms copying src/confluent_kafka/schema_registry/rules/encryption/azurekms/azure_client.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/rules/encryption/azurekms copying src/confluent_kafka/schema_registry/rules/encryption/azurekms/azure_driver.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/rules/encryption/azurekms copying src/confluent_kafka/schema_registry/rules/encryption/azurekms/__init__.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/rules/encryption/azurekms creating build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/rules/encryption/dek_registry copying src/confluent_kafka/schema_registry/rules/encryption/dek_registry/mock_dek_registry_client.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/rules/encryption/dek_registry copying src/confluent_kafka/schema_registry/rules/encryption/dek_registry/dek_registry_client.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/rules/encryption/dek_registry copying src/confluent_kafka/schema_registry/rules/encryption/dek_registry/__init__.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/rules/encryption/dek_registry creating build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/confluent/types copying src/confluent_kafka/schema_registry/confluent/types/decimal_pb2.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/confluent/types copying src/confluent_kafka/schema_registry/confluent/types/__init__.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/confluent/types creating build/lib.linux-x86_64-cpython-39/confluent_kafka/experimental/aio copying src/confluent_kafka/experimental/aio/_AIOConsumer.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/experimental/aio copying src/confluent_kafka/experimental/aio/__init__.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/experimental/aio copying src/confluent_kafka/experimental/aio/_common.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/experimental/aio creating build/lib.linux-x86_64-cpython-39/confluent_kafka/experimental/aio/producer copying src/confluent_kafka/experimental/aio/producer/_AIOProducer.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/experimental/aio/producer copying src/confluent_kafka/experimental/aio/producer/_producer_batch_processor.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/experimental/aio/producer copying src/confluent_kafka/experimental/aio/producer/_kafka_batch_executor.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/experimental/aio/producer copying src/confluent_kafka/experimental/aio/producer/__init__.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/experimental/aio/producer copying src/confluent_kafka/experimental/aio/producer/_buffer_timeout_manager.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/experimental/aio/producer copying src/confluent_kafka/experimental/aio/producer/_message_batch.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/experimental/aio/producer creating build/lib.linux-x86_64-cpython-39/confluent_kafka/avro/serializer copying src/confluent_kafka/avro/serializer/message_serializer.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/avro/serializer copying src/confluent_kafka/avro/serializer/__init__.py -> build/lib.linux-x86_64-cpython-39/confluent_kafka/avro/serializer running build_ext building 'confluent_kafka.cimpl' extension creating build/temp.linux-x86_64-cpython-39 creating build/temp.linux-x86_64-cpython-39/builddir creating build/temp.linux-x86_64-cpython-39/builddir/build creating build/temp.linux-x86_64-cpython-39/builddir/build/BUILD creating build/temp.linux-x86_64-cpython-39/builddir/build/BUILD/confluent_kafka-2.12.0 creating build/temp.linux-x86_64-cpython-39/builddir/build/BUILD/confluent_kafka-2.12.0/src creating build/temp.linux-x86_64-cpython-39/builddir/build/BUILD/confluent_kafka-2.12.0/src/confluent_kafka creating build/temp.linux-x86_64-cpython-39/builddir/build/BUILD/confluent_kafka-2.12.0/src/confluent_kafka/src gcc -Wno-unused-result -Wsign-compare -DDYNAMIC_ANNOTATIONS_ENABLED=1 -DNDEBUG -O2 -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -fstack-protector-strong -m64 -march=x86-64-v2 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -D_GNU_SOURCE -fPIC -fwrapv -O2 -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -fstack-protector-strong -m64 -march=x86-64-v2 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -D_GNU_SOURCE -fPIC -fwrapv -O2 -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -fstack-protector-strong -m64 -march=x86-64-v2 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -D_GNU_SOURCE -fPIC -fwrapv -O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -march=x86-64-v2 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -fPIC -I/usr/include/python3.9 -c /builddir/build/BUILD/confluent_kafka-2.12.0/src/confluent_kafka/src/Admin.c -o build/temp.linux-x86_64-cpython-39/builddir/build/BUILD/confluent_kafka-2.12.0/src/confluent_kafka/src/Admin.o gcc -Wno-unused-result -Wsign-compare -DDYNAMIC_ANNOTATIONS_ENABLED=1 -DNDEBUG -O2 -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -fstack-protector-strong -m64 -march=x86-64-v2 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -D_GNU_SOURCE -fPIC -fwrapv -O2 -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -fstack-protector-strong -m64 -march=x86-64-v2 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -D_GNU_SOURCE -fPIC -fwrapv -O2 -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -fstack-protector-strong -m64 -march=x86-64-v2 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -D_GNU_SOURCE -fPIC -fwrapv -O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -march=x86-64-v2 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -fPIC -I/usr/include/python3.9 -c /builddir/build/BUILD/confluent_kafka-2.12.0/src/confluent_kafka/src/AdminTypes.c -o build/temp.linux-x86_64-cpython-39/builddir/build/BUILD/confluent_kafka-2.12.0/src/confluent_kafka/src/AdminTypes.o gcc -Wno-unused-result -Wsign-compare -DDYNAMIC_ANNOTATIONS_ENABLED=1 -DNDEBUG -O2 -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -fstack-protector-strong -m64 -march=x86-64-v2 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -D_GNU_SOURCE -fPIC -fwrapv -O2 -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -fstack-protector-strong -m64 -march=x86-64-v2 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -D_GNU_SOURCE -fPIC -fwrapv -O2 -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -fstack-protector-strong -m64 -march=x86-64-v2 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -D_GNU_SOURCE -fPIC -fwrapv -O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -march=x86-64-v2 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -fPIC -I/usr/include/python3.9 -c /builddir/build/BUILD/confluent_kafka-2.12.0/src/confluent_kafka/src/Consumer.c -o build/temp.linux-x86_64-cpython-39/builddir/build/BUILD/confluent_kafka-2.12.0/src/confluent_kafka/src/Consumer.o gcc -Wno-unused-result -Wsign-compare -DDYNAMIC_ANNOTATIONS_ENABLED=1 -DNDEBUG -O2 -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -fstack-protector-strong -m64 -march=x86-64-v2 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -D_GNU_SOURCE -fPIC -fwrapv -O2 -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -fstack-protector-strong -m64 -march=x86-64-v2 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -D_GNU_SOURCE -fPIC -fwrapv -O2 -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -fstack-protector-strong -m64 -march=x86-64-v2 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -D_GNU_SOURCE -fPIC -fwrapv -O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -march=x86-64-v2 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -fPIC -I/usr/include/python3.9 -c /builddir/build/BUILD/confluent_kafka-2.12.0/src/confluent_kafka/src/Metadata.c -o build/temp.linux-x86_64-cpython-39/builddir/build/BUILD/confluent_kafka-2.12.0/src/confluent_kafka/src/Metadata.o gcc -Wno-unused-result -Wsign-compare -DDYNAMIC_ANNOTATIONS_ENABLED=1 -DNDEBUG -O2 -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -fstack-protector-strong -m64 -march=x86-64-v2 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -D_GNU_SOURCE -fPIC -fwrapv -O2 -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -fstack-protector-strong -m64 -march=x86-64-v2 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -D_GNU_SOURCE -fPIC -fwrapv -O2 -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -fstack-protector-strong -m64 -march=x86-64-v2 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -D_GNU_SOURCE -fPIC -fwrapv -O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -march=x86-64-v2 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -fPIC -I/usr/include/python3.9 -c /builddir/build/BUILD/confluent_kafka-2.12.0/src/confluent_kafka/src/Producer.c -o build/temp.linux-x86_64-cpython-39/builddir/build/BUILD/confluent_kafka-2.12.0/src/confluent_kafka/src/Producer.o gcc -Wno-unused-result -Wsign-compare -DDYNAMIC_ANNOTATIONS_ENABLED=1 -DNDEBUG -O2 -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -fstack-protector-strong -m64 -march=x86-64-v2 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -D_GNU_SOURCE -fPIC -fwrapv -O2 -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -fstack-protector-strong -m64 -march=x86-64-v2 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -D_GNU_SOURCE -fPIC -fwrapv -O2 -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -fstack-protector-strong -m64 -march=x86-64-v2 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -D_GNU_SOURCE -fPIC -fwrapv -O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -march=x86-64-v2 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -fPIC -I/usr/include/python3.9 -c /builddir/build/BUILD/confluent_kafka-2.12.0/src/confluent_kafka/src/confluent_kafka.c -o build/temp.linux-x86_64-cpython-39/builddir/build/BUILD/confluent_kafka-2.12.0/src/confluent_kafka/src/confluent_kafka.o gcc -shared -Wl,-z,relro -Wl,--as-needed -Wl,-z,now -g -Wl,-z,relro -Wl,--as-needed -Wl,-z,now -g -Wl,-z,relro -Wl,--as-needed -Wl,-z,now -specs=/usr/lib/rpm/redhat/redhat-hardened-ld -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -march=x86-64-v2 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection build/temp.linux-x86_64-cpython-39/builddir/build/BUILD/confluent_kafka-2.12.0/src/confluent_kafka/src/Admin.o build/temp.linux-x86_64-cpython-39/builddir/build/BUILD/confluent_kafka-2.12.0/src/confluent_kafka/src/AdminTypes.o build/temp.linux-x86_64-cpython-39/builddir/build/BUILD/confluent_kafka-2.12.0/src/confluent_kafka/src/Consumer.o build/temp.linux-x86_64-cpython-39/builddir/build/BUILD/confluent_kafka-2.12.0/src/confluent_kafka/src/Metadata.o build/temp.linux-x86_64-cpython-39/builddir/build/BUILD/confluent_kafka-2.12.0/src/confluent_kafka/src/Producer.o build/temp.linux-x86_64-cpython-39/builddir/build/BUILD/confluent_kafka-2.12.0/src/confluent_kafka/src/confluent_kafka.o -L/usr/lib64 -lrdkafka -o build/lib.linux-x86_64-cpython-39/confluent_kafka/cimpl.cpython-39-x86_64-linux-gnu.so installing to build/bdist.linux-x86_64/wheel running install running install_lib creating build/bdist.linux-x86_64 creating build/bdist.linux-x86_64/wheel creating build/bdist.linux-x86_64/wheel/confluent_kafka creating build/bdist.linux-x86_64/wheel/confluent_kafka/_util copying build/lib.linux-x86_64-cpython-39/confluent_kafka/_util/validation_util.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/_util copying build/lib.linux-x86_64-cpython-39/confluent_kafka/_util/conversion_util.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/_util copying build/lib.linux-x86_64-cpython-39/confluent_kafka/_util/__init__.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/_util creating build/bdist.linux-x86_64/wheel/confluent_kafka/serialization copying build/lib.linux-x86_64-cpython-39/confluent_kafka/serialization/__init__.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/serialization creating build/bdist.linux-x86_64/wheel/confluent_kafka/schema_registry creating build/bdist.linux-x86_64/wheel/confluent_kafka/schema_registry/_sync copying build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/_sync/mock_schema_registry_client.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/schema_registry/_sync copying build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/_sync/schema_registry_client.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/schema_registry/_sync copying build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/_sync/json_schema.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/schema_registry/_sync copying build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/_sync/serde.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/schema_registry/_sync copying build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/_sync/avro.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/schema_registry/_sync copying build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/_sync/protobuf.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/schema_registry/_sync copying build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/_sync/__init__.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/schema_registry/_sync copying build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/rule_registry.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/schema_registry creating build/bdist.linux-x86_64/wheel/confluent_kafka/schema_registry/_async copying build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/_async/mock_schema_registry_client.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/schema_registry/_async copying build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/_async/schema_registry_client.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/schema_registry/_async copying build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/_async/json_schema.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/schema_registry/_async copying build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/_async/serde.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/schema_registry/_async copying build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/_async/avro.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/schema_registry/_async copying build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/_async/protobuf.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/schema_registry/_async copying build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/_async/__init__.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/schema_registry/_async copying build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/schema_registry_client.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/schema_registry copying build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/error.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/schema_registry creating build/bdist.linux-x86_64/wheel/confluent_kafka/schema_registry/rules creating build/bdist.linux-x86_64/wheel/confluent_kafka/schema_registry/rules/jsonata copying build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/rules/jsonata/jsonata_executor.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/schema_registry/rules/jsonata copying build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/rules/jsonata/__init__.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/schema_registry/rules/jsonata creating build/bdist.linux-x86_64/wheel/confluent_kafka/schema_registry/rules/encryption creating build/bdist.linux-x86_64/wheel/confluent_kafka/schema_registry/rules/encryption/gcpkms copying build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/rules/encryption/gcpkms/gcp_driver.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/schema_registry/rules/encryption/gcpkms copying build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/rules/encryption/gcpkms/gcp_client.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/schema_registry/rules/encryption/gcpkms copying build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/rules/encryption/gcpkms/__init__.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/schema_registry/rules/encryption/gcpkms creating build/bdist.linux-x86_64/wheel/confluent_kafka/schema_registry/rules/encryption/hcvault copying build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/rules/encryption/hcvault/hcvault_client.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/schema_registry/rules/encryption/hcvault copying build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/rules/encryption/hcvault/hcvault_driver.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/schema_registry/rules/encryption/hcvault copying build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/rules/encryption/hcvault/__init__.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/schema_registry/rules/encryption/hcvault creating build/bdist.linux-x86_64/wheel/confluent_kafka/schema_registry/rules/encryption/awskms copying build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/rules/encryption/awskms/aws_driver.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/schema_registry/rules/encryption/awskms copying build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/rules/encryption/awskms/__init__.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/schema_registry/rules/encryption/awskms creating build/bdist.linux-x86_64/wheel/confluent_kafka/schema_registry/rules/encryption/localkms copying build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/rules/encryption/localkms/local_driver.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/schema_registry/rules/encryption/localkms copying build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/rules/encryption/localkms/local_client.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/schema_registry/rules/encryption/localkms copying build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/rules/encryption/localkms/__init__.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/schema_registry/rules/encryption/localkms copying build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/rules/encryption/encrypt_executor.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/schema_registry/rules/encryption creating build/bdist.linux-x86_64/wheel/confluent_kafka/schema_registry/rules/encryption/azurekms copying build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/rules/encryption/azurekms/azure_aead.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/schema_registry/rules/encryption/azurekms copying build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/rules/encryption/azurekms/azure_client.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/schema_registry/rules/encryption/azurekms copying build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/rules/encryption/azurekms/azure_driver.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/schema_registry/rules/encryption/azurekms copying build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/rules/encryption/azurekms/__init__.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/schema_registry/rules/encryption/azurekms copying build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/rules/encryption/__init__.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/schema_registry/rules/encryption creating build/bdist.linux-x86_64/wheel/confluent_kafka/schema_registry/rules/encryption/dek_registry copying build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/rules/encryption/dek_registry/mock_dek_registry_client.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/schema_registry/rules/encryption/dek_registry copying build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/rules/encryption/dek_registry/dek_registry_client.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/schema_registry/rules/encryption/dek_registry copying build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/rules/encryption/dek_registry/__init__.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/schema_registry/rules/encryption/dek_registry copying build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/rules/encryption/kms_driver_registry.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/schema_registry/rules/encryption copying build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/rules/__init__.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/schema_registry/rules creating build/bdist.linux-x86_64/wheel/confluent_kafka/schema_registry/rules/cel copying build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/rules/cel/extra_func.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/schema_registry/rules/cel copying build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/rules/cel/string_format.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/schema_registry/rules/cel copying build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/rules/cel/cel_executor.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/schema_registry/rules/cel copying build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/rules/cel/cel_field_presence.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/schema_registry/rules/cel copying build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/rules/cel/constraints.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/schema_registry/rules/cel copying build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/rules/cel/__init__.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/schema_registry/rules/cel copying build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/rules/cel/cel_field_executor.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/schema_registry/rules/cel copying build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/json_schema.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/schema_registry copying build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/serde.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/schema_registry creating build/bdist.linux-x86_64/wheel/confluent_kafka/schema_registry/common copying build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/common/schema_registry_client.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/schema_registry/common copying build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/common/json_schema.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/schema_registry/common copying build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/common/serde.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/schema_registry/common copying build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/common/avro.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/schema_registry/common copying build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/common/protobuf.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/schema_registry/common copying build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/common/__init__.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/schema_registry/common copying build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/avro.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/schema_registry copying build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/protobuf.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/schema_registry copying build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/__init__.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/schema_registry creating build/bdist.linux-x86_64/wheel/confluent_kafka/schema_registry/confluent copying build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/confluent/meta_pb2.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/schema_registry/confluent creating build/bdist.linux-x86_64/wheel/confluent_kafka/schema_registry/confluent/types copying build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/confluent/types/decimal_pb2.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/schema_registry/confluent/types copying build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/confluent/types/__init__.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/schema_registry/confluent/types copying build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/confluent/__init__.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/schema_registry/confluent copying build/lib.linux-x86_64-cpython-39/confluent_kafka/schema_registry/wildcard_matcher.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/schema_registry creating build/bdist.linux-x86_64/wheel/confluent_kafka/_model copying build/lib.linux-x86_64-cpython-39/confluent_kafka/_model/__init__.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/_model copying build/lib.linux-x86_64-cpython-39/confluent_kafka/serializing_producer.py -> build/bdist.linux-x86_64/wheel/confluent_kafka creating build/bdist.linux-x86_64/wheel/confluent_kafka/experimental copying build/lib.linux-x86_64-cpython-39/confluent_kafka/experimental/__init__.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/experimental creating build/bdist.linux-x86_64/wheel/confluent_kafka/experimental/aio copying build/lib.linux-x86_64-cpython-39/confluent_kafka/experimental/aio/_AIOConsumer.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/experimental/aio creating build/bdist.linux-x86_64/wheel/confluent_kafka/experimental/aio/producer copying build/lib.linux-x86_64-cpython-39/confluent_kafka/experimental/aio/producer/_AIOProducer.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/experimental/aio/producer copying build/lib.linux-x86_64-cpython-39/confluent_kafka/experimental/aio/producer/_producer_batch_processor.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/experimental/aio/producer copying build/lib.linux-x86_64-cpython-39/confluent_kafka/experimental/aio/producer/_kafka_batch_executor.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/experimental/aio/producer copying build/lib.linux-x86_64-cpython-39/confluent_kafka/experimental/aio/producer/__init__.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/experimental/aio/producer copying build/lib.linux-x86_64-cpython-39/confluent_kafka/experimental/aio/producer/_buffer_timeout_manager.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/experimental/aio/producer copying build/lib.linux-x86_64-cpython-39/confluent_kafka/experimental/aio/producer/_message_batch.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/experimental/aio/producer copying build/lib.linux-x86_64-cpython-39/confluent_kafka/experimental/aio/__init__.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/experimental/aio copying build/lib.linux-x86_64-cpython-39/confluent_kafka/experimental/aio/_common.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/experimental/aio copying build/lib.linux-x86_64-cpython-39/confluent_kafka/cimpl.cpython-39-x86_64-linux-gnu.so -> build/bdist.linux-x86_64/wheel/confluent_kafka copying build/lib.linux-x86_64-cpython-39/confluent_kafka/error.py -> build/bdist.linux-x86_64/wheel/confluent_kafka creating build/bdist.linux-x86_64/wheel/confluent_kafka/admin copying build/lib.linux-x86_64-cpython-39/confluent_kafka/admin/_scram.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/admin copying build/lib.linux-x86_64-cpython-39/confluent_kafka/admin/_acl.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/admin copying build/lib.linux-x86_64-cpython-39/confluent_kafka/admin/_listoffsets.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/admin copying build/lib.linux-x86_64-cpython-39/confluent_kafka/admin/_resource.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/admin copying build/lib.linux-x86_64-cpython-39/confluent_kafka/admin/_config.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/admin copying build/lib.linux-x86_64-cpython-39/confluent_kafka/admin/_cluster.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/admin copying build/lib.linux-x86_64-cpython-39/confluent_kafka/admin/_group.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/admin copying build/lib.linux-x86_64-cpython-39/confluent_kafka/admin/_records.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/admin copying build/lib.linux-x86_64-cpython-39/confluent_kafka/admin/_topic.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/admin copying build/lib.linux-x86_64-cpython-39/confluent_kafka/admin/__init__.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/admin copying build/lib.linux-x86_64-cpython-39/confluent_kafka/admin/_metadata.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/admin creating build/bdist.linux-x86_64/wheel/confluent_kafka/kafkatest copying build/lib.linux-x86_64-cpython-39/confluent_kafka/kafkatest/verifiable_client.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/kafkatest copying build/lib.linux-x86_64-cpython-39/confluent_kafka/kafkatest/verifiable_consumer.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/kafkatest copying build/lib.linux-x86_64-cpython-39/confluent_kafka/kafkatest/verifiable_producer.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/kafkatest copying build/lib.linux-x86_64-cpython-39/confluent_kafka/kafkatest/__init__.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/kafkatest creating build/bdist.linux-x86_64/wheel/confluent_kafka/avro creating build/bdist.linux-x86_64/wheel/confluent_kafka/avro/serializer copying build/lib.linux-x86_64-cpython-39/confluent_kafka/avro/serializer/message_serializer.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/avro/serializer copying build/lib.linux-x86_64-cpython-39/confluent_kafka/avro/serializer/__init__.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/avro/serializer copying build/lib.linux-x86_64-cpython-39/confluent_kafka/avro/error.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/avro copying build/lib.linux-x86_64-cpython-39/confluent_kafka/avro/load.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/avro copying build/lib.linux-x86_64-cpython-39/confluent_kafka/avro/cached_schema_registry_client.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/avro copying build/lib.linux-x86_64-cpython-39/confluent_kafka/avro/__init__.py -> build/bdist.linux-x86_64/wheel/confluent_kafka/avro copying build/lib.linux-x86_64-cpython-39/confluent_kafka/__init__.py -> build/bdist.linux-x86_64/wheel/confluent_kafka copying build/lib.linux-x86_64-cpython-39/confluent_kafka/deserializing_consumer.py -> build/bdist.linux-x86_64/wheel/confluent_kafka running install_egg_info running egg_info writing src/confluent_kafka.egg-info/PKG-INFO writing dependency_links to src/confluent_kafka.egg-info/dependency_links.txt writing requirements to src/confluent_kafka.egg-info/requires.txt writing top-level names to src/confluent_kafka.egg-info/top_level.txt reading manifest file 'src/confluent_kafka.egg-info/SOURCES.txt' reading manifest template 'MANIFEST.in' no previously-included directories found matching 'tests' no previously-included directories found matching 'docs' adding license file 'LICENSE' writing manifest file 'src/confluent_kafka.egg-info/SOURCES.txt' Copying src/confluent_kafka.egg-info to build/bdist.linux-x86_64/wheel/confluent_kafka-2.12.0-py3.9.egg-info running install_scripts adding license file "LICENSE" (matched pattern "LICEN[CS]E*") creating build/bdist.linux-x86_64/wheel/confluent_kafka-2.12.0.dist-info/WHEEL creating '/builddir/build/BUILD/confluent_kafka-2.12.0/.pyproject-builddir/pip-wheel-fdnc2f46/tmpa2vl9iqt/confluent_kafka-2.12.0-cp39-cp39-linux_x86_64.whl' and adding 'build/bdist.linux-x86_64/wheel' to it adding 'confluent_kafka/__init__.py' adding 'confluent_kafka/cimpl.cpython-39-x86_64-linux-gnu.so' adding 'confluent_kafka/deserializing_consumer.py' adding 'confluent_kafka/error.py' adding 'confluent_kafka/serializing_producer.py' adding 'confluent_kafka/_model/__init__.py' adding 'confluent_kafka/_util/__init__.py' adding 'confluent_kafka/_util/conversion_util.py' adding 'confluent_kafka/_util/validation_util.py' adding 'confluent_kafka/admin/__init__.py' adding 'confluent_kafka/admin/_acl.py' adding 'confluent_kafka/admin/_cluster.py' adding 'confluent_kafka/admin/_config.py' adding 'confluent_kafka/admin/_group.py' adding 'confluent_kafka/admin/_listoffsets.py' adding 'confluent_kafka/admin/_metadata.py' adding 'confluent_kafka/admin/_records.py' adding 'confluent_kafka/admin/_resource.py' adding 'confluent_kafka/admin/_scram.py' adding 'confluent_kafka/admin/_topic.py' adding 'confluent_kafka/avro/__init__.py' adding 'confluent_kafka/avro/cached_schema_registry_client.py' adding 'confluent_kafka/avro/error.py' adding 'confluent_kafka/avro/load.py' adding 'confluent_kafka/avro/serializer/__init__.py' adding 'confluent_kafka/avro/serializer/message_serializer.py' adding 'confluent_kafka/experimental/__init__.py' adding 'confluent_kafka/experimental/aio/_AIOConsumer.py' adding 'confluent_kafka/experimental/aio/__init__.py' adding 'confluent_kafka/experimental/aio/_common.py' adding 'confluent_kafka/experimental/aio/producer/_AIOProducer.py' adding 'confluent_kafka/experimental/aio/producer/__init__.py' adding 'confluent_kafka/experimental/aio/producer/_buffer_timeout_manager.py' adding 'confluent_kafka/experimental/aio/producer/_kafka_batch_executor.py' adding 'confluent_kafka/experimental/aio/producer/_message_batch.py' adding 'confluent_kafka/experimental/aio/producer/_producer_batch_processor.py' adding 'confluent_kafka/kafkatest/__init__.py' adding 'confluent_kafka/kafkatest/verifiable_client.py' adding 'confluent_kafka/kafkatest/verifiable_consumer.py' adding 'confluent_kafka/kafkatest/verifiable_producer.py' adding 'confluent_kafka/schema_registry/__init__.py' adding 'confluent_kafka/schema_registry/avro.py' adding 'confluent_kafka/schema_registry/error.py' adding 'confluent_kafka/schema_registry/json_schema.py' adding 'confluent_kafka/schema_registry/protobuf.py' adding 'confluent_kafka/schema_registry/rule_registry.py' adding 'confluent_kafka/schema_registry/schema_registry_client.py' adding 'confluent_kafka/schema_registry/serde.py' adding 'confluent_kafka/schema_registry/wildcard_matcher.py' adding 'confluent_kafka/schema_registry/_async/__init__.py' adding 'confluent_kafka/schema_registry/_async/avro.py' adding 'confluent_kafka/schema_registry/_async/json_schema.py' adding 'confluent_kafka/schema_registry/_async/mock_schema_registry_client.py' adding 'confluent_kafka/schema_registry/_async/protobuf.py' adding 'confluent_kafka/schema_registry/_async/schema_registry_client.py' adding 'confluent_kafka/schema_registry/_async/serde.py' adding 'confluent_kafka/schema_registry/_sync/__init__.py' adding 'confluent_kafka/schema_registry/_sync/avro.py' adding 'confluent_kafka/schema_registry/_sync/json_schema.py' adding 'confluent_kafka/schema_registry/_sync/mock_schema_registry_client.py' adding 'confluent_kafka/schema_registry/_sync/protobuf.py' adding 'confluent_kafka/schema_registry/_sync/schema_registry_client.py' adding 'confluent_kafka/schema_registry/_sync/serde.py' adding 'confluent_kafka/schema_registry/common/__init__.py' adding 'confluent_kafka/schema_registry/common/avro.py' adding 'confluent_kafka/schema_registry/common/json_schema.py' adding 'confluent_kafka/schema_registry/common/protobuf.py' adding 'confluent_kafka/schema_registry/common/schema_registry_client.py' adding 'confluent_kafka/schema_registry/common/serde.py' adding 'confluent_kafka/schema_registry/confluent/__init__.py' adding 'confluent_kafka/schema_registry/confluent/meta_pb2.py' adding 'confluent_kafka/schema_registry/confluent/types/__init__.py' adding 'confluent_kafka/schema_registry/confluent/types/decimal_pb2.py' adding 'confluent_kafka/schema_registry/rules/__init__.py' adding 'confluent_kafka/schema_registry/rules/cel/__init__.py' adding 'confluent_kafka/schema_registry/rules/cel/cel_executor.py' adding 'confluent_kafka/schema_registry/rules/cel/cel_field_executor.py' adding 'confluent_kafka/schema_registry/rules/cel/cel_field_presence.py' adding 'confluent_kafka/schema_registry/rules/cel/constraints.py' adding 'confluent_kafka/schema_registry/rules/cel/extra_func.py' adding 'confluent_kafka/schema_registry/rules/cel/string_format.py' adding 'confluent_kafka/schema_registry/rules/encryption/__init__.py' adding 'confluent_kafka/schema_registry/rules/encryption/encrypt_executor.py' adding 'confluent_kafka/schema_registry/rules/encryption/kms_driver_registry.py' adding 'confluent_kafka/schema_registry/rules/encryption/awskms/__init__.py' adding 'confluent_kafka/schema_registry/rules/encryption/awskms/aws_driver.py' adding 'confluent_kafka/schema_registry/rules/encryption/azurekms/__init__.py' adding 'confluent_kafka/schema_registry/rules/encryption/azurekms/azure_aead.py' adding 'confluent_kafka/schema_registry/rules/encryption/azurekms/azure_client.py' adding 'confluent_kafka/schema_registry/rules/encryption/azurekms/azure_driver.py' adding 'confluent_kafka/schema_registry/rules/encryption/dek_registry/__init__.py' adding 'confluent_kafka/schema_registry/rules/encryption/dek_registry/dek_registry_client.py' adding 'confluent_kafka/schema_registry/rules/encryption/dek_registry/mock_dek_registry_client.py' adding 'confluent_kafka/schema_registry/rules/encryption/gcpkms/__init__.py' adding 'confluent_kafka/schema_registry/rules/encryption/gcpkms/gcp_client.py' adding 'confluent_kafka/schema_registry/rules/encryption/gcpkms/gcp_driver.py' adding 'confluent_kafka/schema_registry/rules/encryption/hcvault/__init__.py' adding 'confluent_kafka/schema_registry/rules/encryption/hcvault/hcvault_client.py' adding 'confluent_kafka/schema_registry/rules/encryption/hcvault/hcvault_driver.py' adding 'confluent_kafka/schema_registry/rules/encryption/localkms/__init__.py' adding 'confluent_kafka/schema_registry/rules/encryption/localkms/local_client.py' adding 'confluent_kafka/schema_registry/rules/encryption/localkms/local_driver.py' adding 'confluent_kafka/schema_registry/rules/jsonata/__init__.py' adding 'confluent_kafka/schema_registry/rules/jsonata/jsonata_executor.py' adding 'confluent_kafka/serialization/__init__.py' adding 'confluent_kafka-2.12.0.dist-info/LICENSE' adding 'confluent_kafka-2.12.0.dist-info/METADATA' adding 'confluent_kafka-2.12.0.dist-info/WHEEL' adding 'confluent_kafka-2.12.0.dist-info/top_level.txt' adding 'confluent_kafka-2.12.0.dist-info/RECORD' removing build/bdist.linux-x86_64/wheel Building wheel for confluent-kafka (pyproject.toml): finished with status 'done' Created wheel for confluent-kafka: filename=confluent_kafka-2.12.0-cp39-cp39-linux_x86_64.whl size=511399 sha256=69b453b55b9d51d7119ed047ae58610c0318b4659635617b1109cb855fa54761 Stored in directory: /builddir/.cache/pip/wheels/c0/27/2a/f586d73d252f056d41e2c307bdca67f90179369a5ac953064e Successfully built confluent-kafka + RPM_EC=0 ++ jobs -p + exit 0 Executing(%install): /bin/sh -e /var/tmp/rpm-tmp.v6oloj + umask 022 + cd /builddir/build/BUILD + '[' /builddir/build/BUILDROOT/python-confluent-kafka-2.12.0-1.el9.x86_64 '!=' / ']' + rm -rf /builddir/build/BUILDROOT/python-confluent-kafka-2.12.0-1.el9.x86_64 ++ dirname /builddir/build/BUILDROOT/python-confluent-kafka-2.12.0-1.el9.x86_64 + mkdir -p /builddir/build/BUILDROOT + mkdir /builddir/build/BUILDROOT/python-confluent-kafka-2.12.0-1.el9.x86_64 + cd confluent_kafka-2.12.0 ++ ls /builddir/build/BUILD/confluent_kafka-2.12.0/pyproject-wheeldir/confluent_kafka-2.12.0-cp39-cp39-linux_x86_64.whl ++ xargs basename --multiple ++ sed -E 's/([^-]+)-([^-]+)-.+\.whl/\1==\2/' + specifier=confluent_kafka==2.12.0 + '[' -z confluent_kafka==2.12.0 ']' + TMPDIR=/builddir/build/BUILD/confluent_kafka-2.12.0/.pyproject-builddir + /usr/bin/python3 -m pip install --root /builddir/build/BUILDROOT/python-confluent-kafka-2.12.0-1.el9.x86_64 --prefix /usr --no-deps --disable-pip-version-check --progress-bar off --verbose --ignore-installed --no-warn-script-location --no-index --no-cache-dir --find-links /builddir/build/BUILD/confluent_kafka-2.12.0/pyproject-wheeldir confluent_kafka==2.12.0 Using pip 21.3.1 from /usr/lib/python3.9/site-packages/pip (python 3.9) Looking in links: /builddir/build/BUILD/confluent_kafka-2.12.0/pyproject-wheeldir Processing ./pyproject-wheeldir/confluent_kafka-2.12.0-cp39-cp39-linux_x86_64.whl Installing collected packages: confluent-kafka Successfully installed confluent-kafka-2.12.0 + '[' -d /builddir/build/BUILDROOT/python-confluent-kafka-2.12.0-1.el9.x86_64/usr/bin ']' + rm -f /builddir/build/BUILD/python-confluent-kafka-2.12.0-1.el9.x86_64-pyproject-ghost-distinfo + site_dirs=() + '[' -d /builddir/build/BUILDROOT/python-confluent-kafka-2.12.0-1.el9.x86_64/usr/lib/python3.9/site-packages ']' + '[' /builddir/build/BUILDROOT/python-confluent-kafka-2.12.0-1.el9.x86_64/usr/lib64/python3.9/site-packages '!=' /builddir/build/BUILDROOT/python-confluent-kafka-2.12.0-1.el9.x86_64/usr/lib/python3.9/site-packages ']' + '[' -d /builddir/build/BUILDROOT/python-confluent-kafka-2.12.0-1.el9.x86_64/usr/lib64/python3.9/site-packages ']' + site_dirs+=("/usr/lib64/python3.9/site-packages") + for site_dir in ${site_dirs[@]} + for distinfo in /builddir/build/BUILDROOT/python-confluent-kafka-2.12.0-1.el9.x86_64$site_dir/*.dist-info + echo '%ghost /usr/lib64/python3.9/site-packages/confluent_kafka-2.12.0.dist-info' + sed -i s/pip/rpm/ /builddir/build/BUILDROOT/python-confluent-kafka-2.12.0-1.el9.x86_64/usr/lib64/python3.9/site-packages/confluent_kafka-2.12.0.dist-info/INSTALLER + PYTHONPATH=/usr/lib/rpm/redhat + /usr/bin/python3 -B /usr/lib/rpm/redhat/pyproject_preprocess_record.py --buildroot /builddir/build/BUILDROOT/python-confluent-kafka-2.12.0-1.el9.x86_64 --record /builddir/build/BUILDROOT/python-confluent-kafka-2.12.0-1.el9.x86_64/usr/lib64/python3.9/site-packages/confluent_kafka-2.12.0.dist-info/RECORD --output /builddir/build/BUILD/python-confluent-kafka-2.12.0-1.el9.x86_64-pyproject-record + rm -fv /builddir/build/BUILDROOT/python-confluent-kafka-2.12.0-1.el9.x86_64/usr/lib64/python3.9/site-packages/confluent_kafka-2.12.0.dist-info/RECORD removed '/builddir/build/BUILDROOT/python-confluent-kafka-2.12.0-1.el9.x86_64/usr/lib64/python3.9/site-packages/confluent_kafka-2.12.0.dist-info/RECORD' + rm -fv /builddir/build/BUILDROOT/python-confluent-kafka-2.12.0-1.el9.x86_64/usr/lib64/python3.9/site-packages/confluent_kafka-2.12.0.dist-info/REQUESTED removed '/builddir/build/BUILDROOT/python-confluent-kafka-2.12.0-1.el9.x86_64/usr/lib64/python3.9/site-packages/confluent_kafka-2.12.0.dist-info/REQUESTED' ++ wc -l /builddir/build/BUILD/python-confluent-kafka-2.12.0-1.el9.x86_64-pyproject-ghost-distinfo ++ cut -f1 '-d ' + lines=1 + '[' 1 -ne 1 ']' + RPM_FILES_ESCAPE=4.18 + /usr/bin/python3 /usr/lib/rpm/redhat/pyproject_save_files.py --output-files /builddir/build/BUILD/python-confluent-kafka-2.12.0-1.el9.x86_64-pyproject-files --output-modules /builddir/build/BUILD/python-confluent-kafka-2.12.0-1.el9.x86_64-pyproject-modules --buildroot /builddir/build/BUILDROOT/python-confluent-kafka-2.12.0-1.el9.x86_64 --sitelib /usr/lib/python3.9/site-packages --sitearch /usr/lib64/python3.9/site-packages --python-version 3.9 --pyproject-record /builddir/build/BUILD/python-confluent-kafka-2.12.0-1.el9.x86_64-pyproject-record --prefix /usr -l confluent_kafka + /usr/bin/find-debuginfo -j8 --strict-build-id -m -i --build-id-seed 2.12.0-1.el9 --unique-debug-suffix -2.12.0-1.el9.x86_64 --unique-debug-src-base python-confluent-kafka-2.12.0-1.el9.x86_64 --run-dwz --dwz-low-mem-die-limit 10000000 --dwz-max-die-limit 110000000 --remove-section .gnu.build.attributes -S debugsourcefiles.list /builddir/build/BUILD/confluent_kafka-2.12.0 extracting debug info from /builddir/build/BUILDROOT/python-confluent-kafka-2.12.0-1.el9.x86_64/usr/lib64/python3.9/site-packages/confluent_kafka/cimpl.cpython-39-x86_64-linux-gnu.so original debug info size: 572kB, size after compression: 532kB /usr/bin/sepdebugcrcfix: Updated 1 CRC32s, 0 CRC32s did match. 913 blocks + /usr/lib/rpm/check-buildroot + /usr/lib/rpm/redhat/brp-ldconfig + /usr/lib/rpm/brp-compress + /usr/lib/rpm/redhat/brp-strip-lto /usr/bin/strip + /usr/lib/rpm/brp-strip-static-archive /usr/bin/strip + /usr/lib/rpm/redhat/brp-python-bytecompile '' 1 0 Bytecompiling .py files below /builddir/build/BUILDROOT/python-confluent-kafka-2.12.0-1.el9.x86_64/usr/lib64/python3.9 using python3.9 Bytecompiling .py files below /builddir/build/BUILDROOT/python-confluent-kafka-2.12.0-1.el9.x86_64/usr/lib/debug/usr/lib64/python3.9 using python3.9 + /usr/lib/rpm/brp-python-hardlink + /usr/lib/rpm/redhat/brp-mangle-shebangs Executing(%check): /bin/sh -e /var/tmp/rpm-tmp.etr5Oz + umask 022 + cd /builddir/build/BUILD + cd confluent_kafka-2.12.0 + PATH=/builddir/build/BUILDROOT/python-confluent-kafka-2.12.0-1.el9.x86_64/usr/bin:/usr/bin:/bin:/usr/sbin:/sbin:/usr/local/sbin + PYTHONPATH=/builddir/build/BUILDROOT/python-confluent-kafka-2.12.0-1.el9.x86_64/usr/lib64/python3.9/site-packages:/builddir/build/BUILDROOT/python-confluent-kafka-2.12.0-1.el9.x86_64/usr/lib/python3.9/site-packages + _PYTHONSITE=/builddir/build/BUILDROOT/python-confluent-kafka-2.12.0-1.el9.x86_64/usr/lib64/python3.9/site-packages:/builddir/build/BUILDROOT/python-confluent-kafka-2.12.0-1.el9.x86_64/usr/lib/python3.9/site-packages + PYTHONDONTWRITEBYTECODE=1 + /usr/bin/python3 -s /usr/lib/rpm/redhat/import_all_modules.py confluent_kafka Check import: confluent_kafka + RPM_EC=0 ++ jobs -p + exit 0 Processing files: python3-confluent-kafka-2.12.0-1.el9.x86_64 Executing(%doc): /bin/sh -e /var/tmp/rpm-tmp.EMzeN8 + umask 022 + cd /builddir/build/BUILD + cd confluent_kafka-2.12.0 + DOCDIR=/builddir/build/BUILDROOT/python-confluent-kafka-2.12.0-1.el9.x86_64/usr/share/doc/python3-confluent-kafka + export LC_ALL=C + LC_ALL=C + export DOCDIR + /usr/bin/mkdir -p /builddir/build/BUILDROOT/python-confluent-kafka-2.12.0-1.el9.x86_64/usr/share/doc/python3-confluent-kafka + cp -pr README.md /builddir/build/BUILDROOT/python-confluent-kafka-2.12.0-1.el9.x86_64/usr/share/doc/python3-confluent-kafka + RPM_EC=0 ++ jobs -p + exit 0 Executing(%license): /bin/sh -e /var/tmp/rpm-tmp.BaEpWA + umask 022 + cd /builddir/build/BUILD + cd confluent_kafka-2.12.0 + LICENSEDIR=/builddir/build/BUILDROOT/python-confluent-kafka-2.12.0-1.el9.x86_64/usr/share/licenses/python3-confluent-kafka + export LC_ALL=C + LC_ALL=C + export LICENSEDIR + /usr/bin/mkdir -p /builddir/build/BUILDROOT/python-confluent-kafka-2.12.0-1.el9.x86_64/usr/share/licenses/python3-confluent-kafka + cp -pr LICENSE /builddir/build/BUILDROOT/python-confluent-kafka-2.12.0-1.el9.x86_64/usr/share/licenses/python3-confluent-kafka + RPM_EC=0 ++ jobs -p + exit 0 Provides: python-confluent-kafka = 2.12.0-1.el9 python3-confluent-kafka = 2.12.0-1.el9 python3-confluent-kafka(x86-64) = 2.12.0-1.el9 python3.9-confluent-kafka = 2.12.0-1.el9 python3.9dist(confluent-kafka) = 2.12 python3dist(confluent-kafka) = 2.12 Requires(rpmlib): rpmlib(CompressedFileNames) <= 3.0.4-1 rpmlib(FileDigests) <= 4.6.0-1 rpmlib(PartialHardlinkSets) <= 4.0.4-1 rpmlib(PayloadFilesHavePrefix) <= 4.0-1 Requires: libc.so.6()(64bit) libc.so.6(GLIBC_2.14)(64bit) libc.so.6(GLIBC_2.2.5)(64bit) libc.so.6(GLIBC_2.3.4)(64bit) libc.so.6(GLIBC_2.4)(64bit) librdkafka.so.1()(64bit) python(abi) = 3.9 python3.9dist(tomli) rtld(GNU_HASH) Obsoletes: python39-confluent-kafka < 2.12.0-1.el9 Processing files: python-confluent-kafka-debugsource-2.12.0-1.el9.x86_64 Provides: python-confluent-kafka-debugsource = 2.12.0-1.el9 python-confluent-kafka-debugsource(x86-64) = 2.12.0-1.el9 Requires(rpmlib): rpmlib(CompressedFileNames) <= 3.0.4-1 rpmlib(FileDigests) <= 4.6.0-1 rpmlib(PayloadFilesHavePrefix) <= 4.0-1 Processing files: python3-confluent-kafka-debuginfo-2.12.0-1.el9.x86_64 Provides: debuginfo(build-id) = cfbdae194666ec9f3c3f15b0b83a17982fe515d4 python-confluent-kafka-debuginfo = 2.12.0-1.el9 python3-confluent-kafka-debuginfo = 2.12.0-1.el9 python3-confluent-kafka-debuginfo(x86-64) = 2.12.0-1.el9 python3.9-confluent-kafka-debuginfo = 2.12.0-1.el9 Requires(rpmlib): rpmlib(CompressedFileNames) <= 3.0.4-1 rpmlib(FileDigests) <= 4.6.0-1 rpmlib(PayloadFilesHavePrefix) <= 4.0-1 Recommends: python-confluent-kafka-debugsource(x86-64) = 2.12.0-1.el9 Checking for unpackaged file(s): /usr/lib/rpm/check-files /builddir/build/BUILDROOT/python-confluent-kafka-2.12.0-1.el9.x86_64 Wrote: /builddir/build/SRPMS/python-confluent-kafka-2.12.0-1.el9.src.rpm Wrote: /builddir/build/RPMS/python3-confluent-kafka-debuginfo-2.12.0-1.el9.x86_64.rpm Wrote: /builddir/build/RPMS/python-confluent-kafka-debugsource-2.12.0-1.el9.x86_64.rpm Wrote: /builddir/build/RPMS/python3-confluent-kafka-2.12.0-1.el9.x86_64.rpm Child return code was: 0