Mock Version: 2.15 ENTER ['do_with_status'](['bash', '--login', '-c', '/usr/bin/rpmbuild -bs --target x86_64 --nodeps /builddir/build/SPECS/python-jsonpath-ng.spec'], chrootPath='/var/lib/mock/fedora-rawhide-x86_64-1642774574.149928/root'env={'TERM': 'vt100', 'SHELL': '/bin/bash', 'HOME': '/builddir', 'HOSTNAME': 'mock', 'PATH': '/usr/bin:/bin:/usr/sbin:/sbin', 'PROMPT_COMMAND': 'printf "\\033]0;\\007"', 'PS1': ' \\s-\\v\\$ ', 'LANG': 'C.UTF-8'}shell=Falselogger=timeout=0uid=1000gid=135user='mockbuild'nspawn_args=['--capability=cap_ipc_lock', '--rlimit=RLIMIT_NOFILE=10240', '--capability=cap_ipc_lock', '--bind=/tmp/mock-resolv.2gtv9nyl:/etc/resolv.conf', '--bind=/dev/btrfs-control', '--bind=/dev/loop-control', '--bind=/dev/loop0', '--bind=/dev/loop1', '--bind=/dev/loop2', '--bind=/dev/loop3', '--bind=/dev/loop4', '--bind=/dev/loop5', '--bind=/dev/loop6', '--bind=/dev/loop7', '--bind=/dev/loop8', '--bind=/dev/loop9', '--bind=/dev/loop10', '--bind=/dev/loop11']unshare_net=TrueprintOutput=True) Using nspawn with args ['--capability=cap_ipc_lock', '--rlimit=RLIMIT_NOFILE=10240', '--capability=cap_ipc_lock', '--bind=/tmp/mock-resolv.2gtv9nyl:/etc/resolv.conf', '--bind=/dev/btrfs-control', '--bind=/dev/loop-control', '--bind=/dev/loop0', '--bind=/dev/loop1', '--bind=/dev/loop2', '--bind=/dev/loop3', '--bind=/dev/loop4', '--bind=/dev/loop5', '--bind=/dev/loop6', '--bind=/dev/loop7', '--bind=/dev/loop8', '--bind=/dev/loop9', '--bind=/dev/loop10', '--bind=/dev/loop11'] Executing command: ['/usr/bin/systemd-nspawn', '-q', '-M', '06ad7fa5f4e64bb392560c1e214e479b', '-D', '/var/lib/mock/fedora-rawhide-x86_64-1642774574.149928/root', '-a', '-u', 'mockbuild', '--capability=cap_ipc_lock', '--rlimit=RLIMIT_NOFILE=10240', '--capability=cap_ipc_lock', '--bind=/tmp/mock-resolv.2gtv9nyl:/etc/resolv.conf', '--bind=/dev/btrfs-control', '--bind=/dev/loop-control', '--bind=/dev/loop0', '--bind=/dev/loop1', '--bind=/dev/loop2', '--bind=/dev/loop3', '--bind=/dev/loop4', '--bind=/dev/loop5', '--bind=/dev/loop6', '--bind=/dev/loop7', '--bind=/dev/loop8', '--bind=/dev/loop9', '--bind=/dev/loop10', '--bind=/dev/loop11', '--console=pipe', '--setenv=TERM=vt100', '--setenv=SHELL=/bin/bash', '--setenv=HOME=/builddir', '--setenv=HOSTNAME=mock', '--setenv=PATH=/usr/bin:/bin:/usr/sbin:/sbin', '--setenv=PROMPT_COMMAND=printf "\\033]0;\\007"', '--setenv=PS1= \\s-\\v\\$ ', '--setenv=LANG=C.UTF-8', '--resolv-conf=off', 'bash', '--login', '-c', '/usr/bin/rpmbuild -bs --target x86_64 --nodeps /builddir/build/SPECS/python-jsonpath-ng.spec'] with env {'TERM': 'vt100', 'SHELL': '/bin/bash', 'HOME': '/builddir', 'HOSTNAME': 'mock', 'PATH': '/usr/bin:/bin:/usr/sbin:/sbin', 'PROMPT_COMMAND': 'printf "\\033]0;\\007"', 'PS1': ' \\s-\\v\\$ ', 'LANG': 'C.UTF-8', 'SYSTEMD_NSPAWN_TMPFS_TMP': '0'} and shell False Building target platforms: x86_64 Building for target x86_64 setting SOURCE_DATE_EPOCH=1642723200 Wrote: /builddir/build/SRPMS/python-jsonpath-ng-1.5.1-6.fc36.src.rpm Child return code was: 0 ENTER ['do_with_status'](['bash', '--login', '-c', '/usr/bin/rpmbuild -bb --target x86_64 --nodeps /builddir/build/SPECS/python-jsonpath-ng.spec'], chrootPath='/var/lib/mock/fedora-rawhide-x86_64-1642774574.149928/root'env={'TERM': 'vt100', 'SHELL': '/bin/bash', 'HOME': '/builddir', 'HOSTNAME': 'mock', 'PATH': '/usr/bin:/bin:/usr/sbin:/sbin', 'PROMPT_COMMAND': 'printf "\\033]0;\\007"', 'PS1': ' \\s-\\v\\$ ', 'LANG': 'C.UTF-8'}shell=Falselogger=timeout=0uid=1000gid=135user='mockbuild'nspawn_args=['--capability=cap_ipc_lock', '--rlimit=RLIMIT_NOFILE=10240', '--capability=cap_ipc_lock', '--bind=/tmp/mock-resolv.2gtv9nyl:/etc/resolv.conf', '--bind=/dev/btrfs-control', '--bind=/dev/loop-control', '--bind=/dev/loop0', '--bind=/dev/loop1', '--bind=/dev/loop2', '--bind=/dev/loop3', '--bind=/dev/loop4', '--bind=/dev/loop5', '--bind=/dev/loop6', '--bind=/dev/loop7', '--bind=/dev/loop8', '--bind=/dev/loop9', '--bind=/dev/loop10', '--bind=/dev/loop11']unshare_net=TrueprintOutput=True) Using nspawn with args ['--capability=cap_ipc_lock', '--rlimit=RLIMIT_NOFILE=10240', '--capability=cap_ipc_lock', '--bind=/tmp/mock-resolv.2gtv9nyl:/etc/resolv.conf', '--bind=/dev/btrfs-control', '--bind=/dev/loop-control', '--bind=/dev/loop0', '--bind=/dev/loop1', '--bind=/dev/loop2', '--bind=/dev/loop3', '--bind=/dev/loop4', '--bind=/dev/loop5', '--bind=/dev/loop6', '--bind=/dev/loop7', '--bind=/dev/loop8', '--bind=/dev/loop9', '--bind=/dev/loop10', '--bind=/dev/loop11'] Executing command: ['/usr/bin/systemd-nspawn', '-q', '-M', '7d2abd07cf9c4b94a695f6fd9a1d6cea', '-D', '/var/lib/mock/fedora-rawhide-x86_64-1642774574.149928/root', '-a', '-u', 'mockbuild', '--capability=cap_ipc_lock', '--rlimit=RLIMIT_NOFILE=10240', '--capability=cap_ipc_lock', '--bind=/tmp/mock-resolv.2gtv9nyl:/etc/resolv.conf', '--bind=/dev/btrfs-control', '--bind=/dev/loop-control', '--bind=/dev/loop0', '--bind=/dev/loop1', '--bind=/dev/loop2', '--bind=/dev/loop3', '--bind=/dev/loop4', '--bind=/dev/loop5', '--bind=/dev/loop6', '--bind=/dev/loop7', '--bind=/dev/loop8', '--bind=/dev/loop9', '--bind=/dev/loop10', '--bind=/dev/loop11', '--console=pipe', '--setenv=TERM=vt100', '--setenv=SHELL=/bin/bash', '--setenv=HOME=/builddir', '--setenv=HOSTNAME=mock', '--setenv=PATH=/usr/bin:/bin:/usr/sbin:/sbin', '--setenv=PROMPT_COMMAND=printf "\\033]0;\\007"', '--setenv=PS1= \\s-\\v\\$ ', '--setenv=LANG=C.UTF-8', '--resolv-conf=off', 'bash', '--login', '-c', '/usr/bin/rpmbuild -bb --target x86_64 --nodeps /builddir/build/SPECS/python-jsonpath-ng.spec'] with env {'TERM': 'vt100', 'SHELL': '/bin/bash', 'HOME': '/builddir', 'HOSTNAME': 'mock', 'PATH': '/usr/bin:/bin:/usr/sbin:/sbin', 'PROMPT_COMMAND': 'printf "\\033]0;\\007"', 'PS1': ' \\s-\\v\\$ ', 'LANG': 'C.UTF-8', 'SYSTEMD_NSPAWN_TMPFS_TMP': '0'} and shell False Building target platforms: x86_64 Building for target x86_64 setting SOURCE_DATE_EPOCH=1642723200 Executing(%prep): /bin/sh -e /var/tmp/rpm-tmp.StK1oG + umask 022 + cd /builddir/build/BUILD + cd /builddir/build/BUILD + rm -rf jsonpath-ng-1.5.1 + /usr/bin/gzip -dc /builddir/build/SOURCES/jsonpath-ng-1.5.1.tar.gz + /usr/bin/tar -xof - + STATUS=0 + '[' 0 -ne 0 ']' + cd jsonpath-ng-1.5.1 + /usr/bin/chmod -Rf a+rX,u+w,g-w,o-w . + sed -i -e '/^#!\//, 1d' jsonpath_ng/bin/jsonpath.py + rm -rf jsonpath-ng.egg-info + RPM_EC=0 ++ jobs -p + exit 0 Executing(%build): /bin/sh -e /var/tmp/rpm-tmp.QWBt0R + umask 022 + cd /builddir/build/BUILD + '[' -f /builddir/build/BUILD/.package_note-python-jsonpath-ng-1.5.1-6.fc36.x86_64.ld ']' + '[' -f /usr/lib/rpm/generate-rpm-note.sh ']' + /usr/lib/rpm/generate-rpm-note.sh python-jsonpath-ng 1.5.1-6.fc36 x86_64 + CFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection' + export CFLAGS + CXXFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection' + export CXXFLAGS + FFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -I/usr/lib64/gfortran/modules' + export FFLAGS + FCFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -I/usr/lib64/gfortran/modules' + export FCFLAGS + LDFLAGS='-Wl,-z,relro -Wl,--as-needed -Wl,-z,now -specs=/usr/lib/rpm/redhat/redhat-hardened-ld -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -Wl,--build-id=sha1 -Wl,-dT,/builddir/build/BUILD/.package_note-python-jsonpath-ng-1.5.1-6.fc36.x86_64.ld' + export LDFLAGS + LT_SYS_LIBRARY_PATH=/usr/lib64: + export LT_SYS_LIBRARY_PATH + CC=gcc + export CC + CXX=g++ + export CXX + cd jsonpath-ng-1.5.1 + CFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection' + LDFLAGS='-Wl,-z,relro -Wl,--as-needed -Wl,-z,now -specs=/usr/lib/rpm/redhat/redhat-hardened-ld -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -Wl,--build-id=sha1 -Wl,-dT,/builddir/build/BUILD/.package_note-python-jsonpath-ng-1.5.1-6.fc36.x86_64.ld' + /usr/bin/python3 setup.py build '--executable=/usr/bin/python3 -s' running build running build_py creating build creating build/lib creating build/lib/jsonpath_ng copying jsonpath_ng/jsonpath.py -> build/lib/jsonpath_ng copying jsonpath_ng/lexer.py -> build/lib/jsonpath_ng copying jsonpath_ng/parser.py -> build/lib/jsonpath_ng copying jsonpath_ng/__init__.py -> build/lib/jsonpath_ng creating build/lib/jsonpath_ng/bin copying jsonpath_ng/bin/jsonpath.py -> build/lib/jsonpath_ng/bin copying jsonpath_ng/bin/__init__.py -> build/lib/jsonpath_ng/bin creating build/lib/jsonpath_ng/ext copying jsonpath_ng/ext/string.py -> build/lib/jsonpath_ng/ext copying jsonpath_ng/ext/filter.py -> build/lib/jsonpath_ng/ext copying jsonpath_ng/ext/parser.py -> build/lib/jsonpath_ng/ext copying jsonpath_ng/ext/__init__.py -> build/lib/jsonpath_ng/ext copying jsonpath_ng/ext/iterable.py -> build/lib/jsonpath_ng/ext copying jsonpath_ng/ext/arithmetic.py -> build/lib/jsonpath_ng/ext + RPM_EC=0 ++ jobs -p + exit 0 Executing(%install): /bin/sh -e /var/tmp/rpm-tmp.dgGtdD + umask 022 + cd /builddir/build/BUILD + '[' /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-6.fc36.x86_64 '!=' / ']' + rm -rf /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-6.fc36.x86_64 ++ dirname /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-6.fc36.x86_64 + mkdir -p /builddir/build/BUILDROOT + mkdir /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-6.fc36.x86_64 + '[' -f /builddir/build/BUILD/.package_note-python-jsonpath-ng-1.5.1-6.fc36.x86_64.ld ']' + CFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection' + export CFLAGS + CXXFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection' + export CXXFLAGS + FFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -I/usr/lib64/gfortran/modules' + export FFLAGS + FCFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -I/usr/lib64/gfortran/modules' + export FCFLAGS + LDFLAGS='-Wl,-z,relro -Wl,--as-needed -Wl,-z,now -specs=/usr/lib/rpm/redhat/redhat-hardened-ld -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -Wl,--build-id=sha1 -Wl,-dT,/builddir/build/BUILD/.package_note-python-jsonpath-ng-1.5.1-6.fc36.x86_64.ld' + export LDFLAGS + LT_SYS_LIBRARY_PATH=/usr/lib64: + export LT_SYS_LIBRARY_PATH + CC=gcc + export CC + CXX=g++ + export CXX + cd jsonpath-ng-1.5.1 + CFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection' + LDFLAGS='-Wl,-z,relro -Wl,--as-needed -Wl,-z,now -specs=/usr/lib/rpm/redhat/redhat-hardened-ld -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -Wl,--build-id=sha1 -Wl,-dT,/builddir/build/BUILD/.package_note-python-jsonpath-ng-1.5.1-6.fc36.x86_64.ld' + /usr/bin/python3 setup.py install -O1 --skip-build --root /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-6.fc36.x86_64 running install /usr/lib/python3.10/site-packages/setuptools/command/install.py:34: SetuptoolsDeprecationWarning: setup.py install is deprecated. Use build and pip and other standards-based tools. warnings.warn( running install_lib creating /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-6.fc36.x86_64/usr creating /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-6.fc36.x86_64/usr/lib creating /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-6.fc36.x86_64/usr/lib/python3.10 creating /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-6.fc36.x86_64/usr/lib/python3.10/site-packages creating /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-6.fc36.x86_64/usr/lib/python3.10/site-packages/jsonpath_ng creating /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-6.fc36.x86_64/usr/lib/python3.10/site-packages/jsonpath_ng/ext copying build/lib/jsonpath_ng/ext/arithmetic.py -> /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-6.fc36.x86_64/usr/lib/python3.10/site-packages/jsonpath_ng/ext copying build/lib/jsonpath_ng/ext/iterable.py -> /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-6.fc36.x86_64/usr/lib/python3.10/site-packages/jsonpath_ng/ext copying build/lib/jsonpath_ng/ext/__init__.py -> /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-6.fc36.x86_64/usr/lib/python3.10/site-packages/jsonpath_ng/ext copying build/lib/jsonpath_ng/ext/parser.py -> /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-6.fc36.x86_64/usr/lib/python3.10/site-packages/jsonpath_ng/ext copying build/lib/jsonpath_ng/ext/filter.py -> /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-6.fc36.x86_64/usr/lib/python3.10/site-packages/jsonpath_ng/ext copying build/lib/jsonpath_ng/ext/string.py -> /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-6.fc36.x86_64/usr/lib/python3.10/site-packages/jsonpath_ng/ext creating /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-6.fc36.x86_64/usr/lib/python3.10/site-packages/jsonpath_ng/bin copying build/lib/jsonpath_ng/bin/__init__.py -> /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-6.fc36.x86_64/usr/lib/python3.10/site-packages/jsonpath_ng/bin copying build/lib/jsonpath_ng/bin/jsonpath.py -> /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-6.fc36.x86_64/usr/lib/python3.10/site-packages/jsonpath_ng/bin copying build/lib/jsonpath_ng/__init__.py -> /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-6.fc36.x86_64/usr/lib/python3.10/site-packages/jsonpath_ng copying build/lib/jsonpath_ng/parser.py -> /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-6.fc36.x86_64/usr/lib/python3.10/site-packages/jsonpath_ng copying build/lib/jsonpath_ng/lexer.py -> /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-6.fc36.x86_64/usr/lib/python3.10/site-packages/jsonpath_ng copying build/lib/jsonpath_ng/jsonpath.py -> /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-6.fc36.x86_64/usr/lib/python3.10/site-packages/jsonpath_ng byte-compiling /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-6.fc36.x86_64/usr/lib/python3.10/site-packages/jsonpath_ng/ext/arithmetic.py to arithmetic.cpython-310.pyc byte-compiling /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-6.fc36.x86_64/usr/lib/python3.10/site-packages/jsonpath_ng/ext/iterable.py to iterable.cpython-310.pyc byte-compiling /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-6.fc36.x86_64/usr/lib/python3.10/site-packages/jsonpath_ng/ext/__init__.py to __init__.cpython-310.pyc byte-compiling /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-6.fc36.x86_64/usr/lib/python3.10/site-packages/jsonpath_ng/ext/parser.py to parser.cpython-310.pyc byte-compiling /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-6.fc36.x86_64/usr/lib/python3.10/site-packages/jsonpath_ng/ext/filter.py to filter.cpython-310.pyc byte-compiling /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-6.fc36.x86_64/usr/lib/python3.10/site-packages/jsonpath_ng/ext/string.py to string.cpython-310.pyc byte-compiling /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-6.fc36.x86_64/usr/lib/python3.10/site-packages/jsonpath_ng/bin/__init__.py to __init__.cpython-310.pyc byte-compiling /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-6.fc36.x86_64/usr/lib/python3.10/site-packages/jsonpath_ng/bin/jsonpath.py to jsonpath.cpython-310.pyc byte-compiling /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-6.fc36.x86_64/usr/lib/python3.10/site-packages/jsonpath_ng/__init__.py to __init__.cpython-310.pyc byte-compiling /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-6.fc36.x86_64/usr/lib/python3.10/site-packages/jsonpath_ng/parser.py to parser.cpython-310.pyc byte-compiling /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-6.fc36.x86_64/usr/lib/python3.10/site-packages/jsonpath_ng/lexer.py to lexer.cpython-310.pyc byte-compiling /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-6.fc36.x86_64/usr/lib/python3.10/site-packages/jsonpath_ng/jsonpath.py to jsonpath.cpython-310.pyc writing byte-compilation script '/tmp/tmpjf_mnzwb.py' /usr/bin/python3 -Wignore:The distutils package is deprecated and slated for removal in Python 3.12. Use setuptools or check PEP 632 for potential alternatives:DeprecationWarning /tmp/tmpjf_mnzwb.py removing /tmp/tmpjf_mnzwb.py running install_egg_info running egg_info writing jsonpath_ng.egg-info/PKG-INFO writing dependency_links to jsonpath_ng.egg-info/dependency_links.txt writing entry points to jsonpath_ng.egg-info/entry_points.txt writing requirements to jsonpath_ng.egg-info/requires.txt writing top-level names to jsonpath_ng.egg-info/top_level.txt reading manifest file 'jsonpath_ng.egg-info/SOURCES.txt' reading manifest template 'MANIFEST.in' writing manifest file 'jsonpath_ng.egg-info/SOURCES.txt' Copying jsonpath_ng.egg-info to /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-6.fc36.x86_64/usr/lib/python3.10/site-packages/jsonpath_ng-1.5.1-py3.10.egg-info running install_scripts Installing jsonpath_ng script to /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-6.fc36.x86_64/usr/bin + rm -rfv /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-6.fc36.x86_64/usr/bin/__pycache__ + /usr/bin/find-debuginfo -j2 --strict-build-id -m -i --build-id-seed 1.5.1-6.fc36 --unique-debug-suffix -1.5.1-6.fc36.x86_64 --unique-debug-src-base python-jsonpath-ng-1.5.1-6.fc36.x86_64 --run-dwz --dwz-low-mem-die-limit 10000000 --dwz-max-die-limit 110000000 -S debugsourcefiles.list /builddir/build/BUILD/jsonpath-ng-1.5.1 find: 'debug': No such file or directory + /usr/lib/rpm/check-buildroot + /usr/lib/rpm/redhat/brp-ldconfig + /usr/lib/rpm/brp-compress + /usr/lib/rpm/redhat/brp-strip-lto /usr/bin/strip + /usr/lib/rpm/brp-strip-static-archive /usr/bin/strip + /usr/lib/rpm/check-rpaths + /usr/lib/rpm/redhat/brp-mangle-shebangs + /usr/lib/rpm/redhat/brp-python-bytecompile '' 1 0 Bytecompiling .py files below /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-6.fc36.x86_64/usr/lib/python3.10 using python3.10 + /usr/lib/rpm/redhat/brp-python-hardlink Executing(%check): /bin/sh -e /var/tmp/rpm-tmp.yKxYcz + umask 022 + cd /builddir/build/BUILD + '[' -f /builddir/build/BUILD/.package_note-python-jsonpath-ng-1.5.1-6.fc36.x86_64.ld ']' + CFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection' + export CFLAGS + CXXFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection' + export CXXFLAGS + FFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -I/usr/lib64/gfortran/modules' + export FFLAGS + FCFLAGS='-O2 -flto=auto -ffat-lto-objects -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -specs=/usr/lib/rpm/redhat/redhat-hardened-cc1 -fstack-protector-strong -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -m64 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -I/usr/lib64/gfortran/modules' + export FCFLAGS + LDFLAGS='-Wl,-z,relro -Wl,--as-needed -Wl,-z,now -specs=/usr/lib/rpm/redhat/redhat-hardened-ld -specs=/usr/lib/rpm/redhat/redhat-annobin-cc1 -Wl,--build-id=sha1 -Wl,-dT,/builddir/build/BUILD/.package_note-python-jsonpath-ng-1.5.1-6.fc36.x86_64.ld' + export LDFLAGS + LT_SYS_LIBRARY_PATH=/usr/lib64: + export LT_SYS_LIBRARY_PATH + CC=gcc + export CC + CXX=g++ + export CXX + cd jsonpath-ng-1.5.1 + /usr/bin/python3 setup.py test running test WARNING: Testing via this command is deprecated and will be removed in a future version. Users looking for a generic test entry point independent of test runner are encouraged to use tox. running egg_info writing jsonpath_ng.egg-info/PKG-INFO writing dependency_links to jsonpath_ng.egg-info/dependency_links.txt writing entry points to jsonpath_ng.egg-info/entry_points.txt writing requirements to jsonpath_ng.egg-info/requires.txt writing top-level names to jsonpath_ng.egg-info/top_level.txt reading manifest file 'jsonpath_ng.egg-info/SOURCES.txt' reading manifest template 'MANIFEST.in' writing manifest file 'jsonpath_ng.egg-info/SOURCES.txt' running build_ext test_DatumInContext_in_context (tests.test_jsonpath.TestDatumInContext) ... ok test_DatumInContext_init (tests.test_jsonpath.TestDatumInContext) ... ok test_child_auto_id (tests.test_jsonpath.TestJsonPath) ... ok test_child_paths (tests.test_jsonpath.TestJsonPath) ... ok test_child_value (tests.test_jsonpath.TestJsonPath) ... ok test_descendants_auto_id (tests.test_jsonpath.TestJsonPath) ... ok test_descendants_paths (tests.test_jsonpath.TestJsonPath) ... ok test_descendants_value (tests.test_jsonpath.TestJsonPath) ... ok test_fields_auto_id (tests.test_jsonpath.TestJsonPath) ... ok test_fields_paths (tests.test_jsonpath.TestJsonPath) ... ok test_fields_value (tests.test_jsonpath.TestJsonPath) ... ok test_hyphen_key (tests.test_jsonpath.TestJsonPath) ... ok test_index_auto_id (tests.test_jsonpath.TestJsonPath) ... ok test_index_paths (tests.test_jsonpath.TestJsonPath) ... ok test_index_value (tests.test_jsonpath.TestJsonPath) ... ok test_parent_value (tests.test_jsonpath.TestJsonPath) ... ok test_root_auto_id (tests.test_jsonpath.TestJsonPath) ... ok test_root_paths (tests.test_jsonpath.TestJsonPath) ... ok test_root_value (tests.test_jsonpath.TestJsonPath) ... ok test_slice_auto_id (tests.test_jsonpath.TestJsonPath) ... ok test_slice_paths (tests.test_jsonpath.TestJsonPath) ... ok test_slice_value (tests.test_jsonpath.TestJsonPath) ... ok test_this_auto_id (tests.test_jsonpath.TestJsonPath) ... ok test_this_paths (tests.test_jsonpath.TestJsonPath) ... ok test_this_value (tests.test_jsonpath.TestJsonPath) ... ok test_update_child (tests.test_jsonpath.TestJsonPath) ... ok test_update_descendants (tests.test_jsonpath.TestJsonPath) ... ok test_update_descendants_where (tests.test_jsonpath.TestJsonPath) ... ok test_update_fields (tests.test_jsonpath.TestJsonPath) ... ok test_update_index (tests.test_jsonpath.TestJsonPath) ... ok test_update_root (tests.test_jsonpath.TestJsonPath) ... ok test_update_slice (tests.test_jsonpath.TestJsonPath) ... ok test_update_this (tests.test_jsonpath.TestJsonPath) ... ok test_update_where (tests.test_jsonpath.TestJsonPath) ... ok test_basic_errors (tests.test_lexer.TestLexer) ... lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' ok test_simple_inputs (tests.test_lexer.TestLexer) ... lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' lex: tokens = ['DOUBLEDOT', 'NUMBER', 'ID', 'NAMED_OPERATOR', 'WHERE'] lex: literals = ['*', '.', '[', ']', '(', ')', '$', ',', ':', '|', '&', '~'] lex: states = {'INITIAL': 'inclusive', 'singlequote': 'exclusive', 'doublequote': 'exclusive', 'backquote': 'exclusive'} lex: Adding rule t_ID -> '[a-zA-Z_@][a-zA-Z0-9_@\-]*' (state 'INITIAL') lex: Adding rule t_NUMBER -> '-?\d+' (state 'INITIAL') lex: Adding rule t_singlequote -> ''' (state 'INITIAL') lex: Adding rule t_doublequote -> '"' (state 'INITIAL') lex: Adding rule t_backquote -> '`' (state 'INITIAL') lex: Adding rule t_newline -> '\n' (state 'INITIAL') lex: Adding rule t_DOUBLEDOT -> '\.\.' (state 'INITIAL') lex: Adding rule t_singlequote_content -> '[^'\\]+' (state 'singlequote') lex: Adding rule t_singlequote_escape -> '\\.' (state 'singlequote') lex: Adding rule t_singlequote_end -> ''' (state 'singlequote') lex: Adding rule t_doublequote_content -> '[^"\\]+' (state 'doublequote') lex: Adding rule t_doublequote_escape -> '\\.' (state 'doublequote') lex: Adding rule t_doublequote_end -> '"' (state 'doublequote') lex: Adding rule t_backquote_escape -> '\\.' (state 'backquote') lex: Adding rule t_backquote_content -> '[^`\\]+' (state 'backquote') lex: Adding rule t_backquote_end -> '`' (state 'backquote') lex: ==== MASTER REGEXS FOLLOW ==== lex: state 'INITIAL' : regex[0] = '(?P[a-zA-Z_@][a-zA-Z0-9_@\-]*)|(?P-?\d+)|(?P')|(?P")|(?P`)|(?P\n)|(?P\.\.)' lex: state 'singlequote' : regex[0] = '(?P[^'\\]+)|(?P\\.)|(?P')' lex: state 'doublequote' : regex[0] = '(?P[^"\\]+)|(?P\\.)|(?P")' lex: state 'backquote' : regex[0] = '(?P\\.)|(?P[^`\\]+)|(?P`)' ok test_filename_mode (tests.bin.test_jsonpath.TestJsonPathScript) ... ok test_stdin_mode (tests.bin.test_jsonpath.TestJsonPathScript) ... ok test_child_auto_id (tests.test_jsonpath_rw_ext.TestJsonPath) tests.test_jsonpath_rw_ext.TestJsonPath.test_child_auto_id ... ok test_child_paths (tests.test_jsonpath_rw_ext.TestJsonPath) tests.test_jsonpath_rw_ext.TestJsonPath.test_child_paths ... ok test_child_value (tests.test_jsonpath_rw_ext.TestJsonPath) tests.test_jsonpath_rw_ext.TestJsonPath.test_child_value ... ok test_descendants_auto_id (tests.test_jsonpath_rw_ext.TestJsonPath) tests.test_jsonpath_rw_ext.TestJsonPath.test_descendants_auto_id ... ok test_descendants_paths (tests.test_jsonpath_rw_ext.TestJsonPath) tests.test_jsonpath_rw_ext.TestJsonPath.test_descendants_paths ... ok test_descendants_value (tests.test_jsonpath_rw_ext.TestJsonPath) tests.test_jsonpath_rw_ext.TestJsonPath.test_descendants_value ... ok test_fields_auto_id (tests.test_jsonpath_rw_ext.TestJsonPath) tests.test_jsonpath_rw_ext.TestJsonPath.test_fields_auto_id ... ok test_fields_paths (tests.test_jsonpath_rw_ext.TestJsonPath) tests.test_jsonpath_rw_ext.TestJsonPath.test_fields_paths ... parse("foo.baz.id").find({'foo': {'baz': 3}}) =?= ['foo.baz'] parse("foo.baz.id").find({'foo': {'baz': [3]}}) =?= ['foo.baz'] parse("foo.baz.id").find({'foo': {'id': 'bizzle', 'baz': 3}}) =?= ['bizzle.baz'] parse("foo.baz.id").find({'foo': {'baz': {'id': 'hi'}}}) =?= ['foo.hi'] parse("foo.baz.bizzle.id").find({'foo': {'baz': {'bizzle': 5}}}) =?= ['foo.baz.bizzle'] parse("foo.baz").find({'foo': {'baz': 3}}).paths =?= ['foo.baz'] parse("foo.baz").find({'foo': {'baz': [3]}}).paths =?= ['foo.baz'] parse("foo.baz.bizzle").find({'foo': {'baz': {'bizzle': 5}}}).paths =?= ['foo.baz.bizzle'] parse("foo.baz").find({'foo': {'baz': 3}}) =?= [3] parse("foo.baz").find({'foo': {'baz': [3]}}) =?= [[3]] parse("foo.baz.bizzle").find({'foo': {'baz': {'bizzle': 5}}}) =?= [5] parse("foo..baz.id").find({'foo': {'baz': 1, 'bing': {'baz': 2}}}) =?= ['foo.baz', 'foo.bing.baz'] parse("foo..baz").find({'foo': {'baz': 1, 'bing': {'baz': 2}}}).paths =?= ['foo.baz', 'foo.bing.baz'] parse("foo..baz").find({'foo': {'baz': 1, 'bing': {'baz': 2}}}) =?= [1, 2] parse("foo..baz").find({'foo': [{'baz': 1}, {'baz': 2}]}) =?= [1, 2] parse("foo.id").find({'foo': 'baz'}) =?= ['foo'] parse("foo.id").find({'foo': {'id': 'baz'}}) =?= ['baz'] parse("foo,baz.id").find({'foo': 1, 'baz': 2}) =?= ['foo', 'baz'] parse("*.id").find({'foo': {'id': 1}, 'baz': 2}) =?= {'baz', '1'} parse("foo").find({'foo': 'baz'}).paths =?= ['foo'] parse("foo,baz").find({'foo': 1, 'baz': 2}).paths =?= ['foo', 'baz'] parse("*").find({'foo': 1, 'baz': 2}).paths =?= {'foo', 'baz'} parse("*").find({'foo': 1, 'baz': 2}).paths =?= {'id', 'foo', 'baz'} parse("foo").find({'foo': 'baz'}) =?= ['baz'] parse("foo,baz").find({'foo': 1, 'baz': 2}) =?= [1, 2] parse("@foo").find({'@foo': 1}) =?= [1] parse("*").find({'foo': 1, 'baz': 2}) =?= {1, 2} parse("*").find({'foo': 1, 'baz': 2}) =?= {1, 2, '`this`'} parse("foo.bar-baz").find({'foo': {'bar-baz': 3}}) =?= [3] parse("foo.[bar-baz,blah-blah]").find({'foo': {'bar-baz': 3, 'blah-blah': 5}}) =?= [3, 5] parse("foo.-baz").find({'foo': {'-baz': 8}}) =?= [8] parse("[0].id").find([42]) =?= ['[0]'] parse("[2].id").find([34, 65, 29, 59]) =?= ['[2]'] parse("[0]").find([42]).paths =?= ['[0]'] parse("[2]").find([34, 65, 29, 59]).paths =?= ['[2]'] parse("[0]").find([42]) =?= [42] parse("[5]").find([42]) =?= [] parse("[2]").find([34, 65, 29, 59]) =?= [29] parse("[0]").find(None) =?= [] parse("foo.baz.`parent`").find({'foo': {'baz': 3}}) =?= [{'baz': 3}] parse("foo.`parent`.foo.baz.`parent`.baz.bizzle").find({'foo': {'baz': {'bizzle': 5}}}) =?= [5] parse("$.id").find({'foo': 'baz'}) =?= ['$'] parse("foo.$.id").find({'foo': 'baz', 'id': 'bizzle'}) =?= ['bizzle'] parse("foo.$.baz.id").find({'foo': 4, 'baz': 3}) =?= ['baz'] parse("$").find({'foo': 'baz'}).paths =?= ['$'] parse("foo.$").find({'foo': 'baz'}).paths =?= ['$'] parse("foo.$.foo").find({'foo': 'baz'}).paths =?= ['foo'] parse("$").find({'foo': 'baz'}) =?= [{'foo': 'baz'}] parse("foo.$").find({'foo': 'baz'}) =?= [{'foo': 'baz'}] parse("foo.$.foo").find({'foo': 'baz'}) =?= ['baz'] parse("[*].id").find([1, 2, 3]) =?= ['[0]', '[1]', '[2]'] parse("[1:].id").find([1, 2, 3, 4]) =?= ['[1]', '[2]', '[3]'] parse("[*]").find([1, 2, 3]).paths =?= ['[0]', '[1]', '[2]'] parse("[1:]").find([1, 2, 3, 4]).paths =?= ['[1]', '[2]', '[3]'] parse("[*]").find([1, 2, 3]) =?= [1, 2, 3] parse("[*]").find(range(1, 4)) =?= [1, 2, 3] parse("[1:]").find([1, 2, 3, 4]) =?= [2, 3, 4] parse("[:2]").find([1, 2, 3, 4]) =?= [1, 2] parse("[*]").find(1) =?= [1] parse("[0:]").find(1) =?= [1] parse("[*]").find({'foo': 1}) =?= [{'foo': 1}] parse("[*].foo").find({'foo': 1}) =?= [1] parse("id").find({'foo': 'baz'}) =?= ['`this`'] parse("foo.`this`.id").find({'foo': 'baz'}) =?= ['foo'] parse("foo.`this`.baz.id").find({'foo': {'baz': 3}}) =?= ['foo.baz'] parse("`this`").find({'foo': 'baz'}).paths =?= ['`this`'] parse("foo.`this`").find({'foo': 'baz'}).paths =?= ['foo'] parse("foo.`this`.baz").find({'foo': {'baz': 3}}).paths =?= ['foo.baz'] parse("`this`").find({'foo': 'baz'}) =?= [{'foo': 'baz'}] parse("foo.`this`").find({'foo': 'baz'}) =?= ['baz'] parse("foo.`this`.baz").find({'foo': {'baz': 3}}) =?= [3] parse('$.foo').update({'foo': 'bar'}, 'baz') =?= {'foo': 'baz'} parse('foo.bar').update({'foo': {'bar': 1}}, 'baz') =?= {'foo': {'bar': 'baz'}} parse('$..somefield').update({'somefield': 1}, 42) =?= {'somefield': 42} parse('$..nestedfield').update({'outer': {'nestedfield': 1}}, 42) =?= {'outer': {'nestedfield': 42}} parse('$..bar').update({'outs': {'bar': 1, 'ins': {'bar': 9}}, 'outs2': {'bar': 2}}, 42) =?= {'outs': {'bar': 42, 'ins': {'bar': 42}}, 'outs2': {'bar': 42}} parse('(* where flag) .. bar').update({'foo': {'bar': 1, 'flag': 1}, 'baz': {'bar': 2}}, 3) =?= {'foo': {'bar': 3, 'flag': 1}, 'baz': {'bar': 2}} parse('foo').update({'foo': 1}, 5) =?= {'foo': 5} parse('$.*').update({'foo': 1, 'bar': 2}, 3) =?= {'foo': 3, 'bar': 3} parse('[0]').update(['foo', 'bar', 'baz'], 'test') =?= ['test', 'bar', 'baz'] parse('$').update('foo', 'bar') =?= 'bar' parse('[0:2]').update(['foo', 'bar', 'baz'], 'test') =?= ['test', 'test', 'baz'] parse('`this`').update('foo', 'bar') =?= 'bar' parse('*.bar where baz').update({'foo': {'bar': {'baz': 1}}, 'bar': {'baz': 2}}, 5) =?= {'foo': {'bar': 5}, 'bar': {'baz': 2}} LexToken($,'$',1,0) LexToken($,'$',-1,-1) LexToken(ID,'hello',1,6) LexToken(ID,'hello',-1,-1) LexToken(ID,'goodbye',1,8) LexToken(ID,'goodbye',-1,-1) LexToken(ID,'doublequote"',1,13) LexToken(ID,'doublequote"',-1,-1) LexToken(ID,'doublequote"',1,14) LexToken(ID,'doublequote"',-1,-1) LexToken(ID,"singlequote'",1,14) LexToken(ID,"singlequote'",-1,-1) LexToken(ID,"singlequote'",1,13) LexToken(ID,"singlequote'",-1,-1) LexToken(ID,'fuzz',1,0) LexToken(ID,'fuzz',-1,-1) LexToken(NUMBER,1,1,0) LexToken(NUMBER,1,-1,-1) LexToken(NUMBER,45,1,0) LexToken(NUMBER,45,-1,-1) LexToken(NUMBER,-1,1,0) LexToken(NUMBER,-1,-1,-1) LexToken(NUMBER,-13,1,1) LexToken(NUMBER,-13,-1,-1) LexToken(ID,'fuzz.bang',1,10) LexToken(ID,'fuzz.bang',-1,-1) LexToken(ID,'fuzz',1,0) LexToken(ID,'fuzz',-1,-1) LexToken(.,'.',1,4) LexToken(.,'.',-1,-1) LexToken(ID,'bang',1,5) LexToken(ID,'bang',-1,-1) LexToken(ID,'fuzz',1,0) LexToken(ID,'fuzz',-1,-1) LexToken(.,'.',1,4) LexToken(.,'.',-1,-1) LexToken(*,'*',1,5) LexToken(*,'*',-1,-1) LexToken(ID,'fuzz',1,0) LexToken(ID,'fuzz',-1,-1) LexToken(DOUBLEDOT,'..',1,4) LexToken(DOUBLEDOT,'..',-1,-1) LexToken(ID,'bang',1,6) LexToken(ID,'bang',-1,-1) LexToken(&,'&',1,0) LexToken(&,'&',-1,-1) LexToken(ID,'@',1,0) LexToken(ID,'@',-1,-1) LexToken(NAMED_OPERATOR,'this',1,5) LexToken(NAMED_OPERATOR,'this',-1,-1) LexToken(|,'|',1,0) LexToken(|,'|',-1,-1) LexToken(WHERE,'where',1,0) LexToken(WHERE,'where',-1,-1) parse("foo.baz.id").find({'foo': {'baz': 3}}) =?= ['foo.baz'] parse("foo.baz.id").find({'foo': {'baz': [3]}}) =?= ['foo.baz'] parse("foo.baz.id").find({'foo': {'id': 'bizzle', 'baz': 3}}) =?= ['bizzle.baz'] parse("foo.baz.id").find({'foo': {'baz': {'id': 'hi'}}}) =?= ['foo.hi'] parse("foo.baz.bizzle.id").find({'foo': {'baz': {'bizzle': 5}}}) =?= ['foo.baz.bizzle'] parse("foo.baz").find({'foo': {'baz': 3}}).paths =?= ['foo.baz'] parse("foo.baz").find({'foo': {'baz': [3]}}).paths =?= ['foo.baz'] parse("foo.baz.bizzle").find({'foo': {'baz': {'bizzle': 5}}}).paths =?= ['foo.baz.bizzle'] parse("foo.baz").find({'foo': {'baz': 3}}) =?= [3] parse("foo.baz").find({'foo': {'baz': [3]}}) =?= [[3]] parse("foo.baz.bizzle").find({'foo': {'baz': {'bizzle': 5}}}) =?= [5] parse("foo..baz.id").find({'foo': {'baz': 1, 'bing': {'baz': 2}}}) =?= ['foo.baz', 'foo.bing.baz'] parse("foo..baz").find({'foo': {'baz': 1, 'bing': {'baz': 2}}}).paths =?= ['foo.baz', 'foo.bing.baz'] parse("foo..baz").find({'foo': {'baz': 1, 'bing': {'baz': 2}}}) =?= [1, 2] parse("foo..baz").find({'foo': [{'baz': 1}, {'baz': 2}]}) =?= [1, 2] parse("foo.id").find({'foo': 'baz'}) =?= ['foo'] parse("foo.id").find({'foo': {'id': 'baz'}}) =?= ['baz'] parse("foo,baz.id").find({'foo': 1, 'baz': 2}) =?= ['foo', 'baz'] parse("*.id").find({'foo': {'id': 1}, 'baz': 2}) =?= {'baz', '1'} parse("foo").find({'foo': 'baz'}).paths =?= ['foo'] parse("foo,baz").find({'foo': 1, 'baz': 2}).paths =?= ['foo', 'baz'] ok test_fields_value (tests.test_jsonpath_rw_ext.TestJsonPath) tests.test_jsonpath_rw_ext.TestJsonPath.test_fields_value ... ok test_hyphen_key (tests.test_jsonpath_rw_ext.TestJsonPath) tests.test_jsonpath_rw_ext.TestJsonPath.test_hyphen_key ... ok test_index_auto_id (tests.test_jsonpath_rw_ext.TestJsonPath) tests.test_jsonpath_rw_ext.TestJsonPath.test_index_auto_id ... ok test_index_paths (tests.test_jsonpath_rw_ext.TestJsonPath) tests.test_jsonpath_rw_ext.TestJsonPath.test_index_paths ... ok test_index_value (tests.test_jsonpath_rw_ext.TestJsonPath) tests.test_jsonpath_rw_ext.TestJsonPath.test_index_value ... ok test_parent_value (tests.test_jsonpath_rw_ext.TestJsonPath) tests.test_jsonpath_rw_ext.TestJsonPath.test_parent_value ... ok test_root_auto_id (tests.test_jsonpath_rw_ext.TestJsonPath) tests.test_jsonpath_rw_ext.TestJsonPath.test_root_auto_id ... ok test_root_paths (tests.test_jsonpath_rw_ext.TestJsonPath) tests.test_jsonpath_rw_ext.TestJsonPath.test_root_paths ... ok test_root_value (tests.test_jsonpath_rw_ext.TestJsonPath) tests.test_jsonpath_rw_ext.TestJsonPath.test_root_value ... ok test_slice_auto_id (tests.test_jsonpath_rw_ext.TestJsonPath) tests.test_jsonpath_rw_ext.TestJsonPath.test_slice_auto_id ... ok test_slice_paths (tests.test_jsonpath_rw_ext.TestJsonPath) tests.test_jsonpath_rw_ext.TestJsonPath.test_slice_paths ... ok test_slice_value (tests.test_jsonpath_rw_ext.TestJsonPath) tests.test_jsonpath_rw_ext.TestJsonPath.test_slice_value ... ok test_this_auto_id (tests.test_jsonpath_rw_ext.TestJsonPath) tests.test_jsonpath_rw_ext.TestJsonPath.test_this_auto_id ... ok test_this_paths (tests.test_jsonpath_rw_ext.TestJsonPath) tests.test_jsonpath_rw_ext.TestJsonPath.test_this_paths ... ok test_this_value (tests.test_jsonpath_rw_ext.TestJsonPath) tests.test_jsonpath_rw_ext.TestJsonPath.test_this_value ... ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (sorted_list) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts /builddir/build/BUILD/jsonpath-ng-1.5.1/jsonpath_ng/parser.py:47: ResourceWarning: unclosed file <_io.TextIOWrapper name='/builddir/build/BUILD/jsonpath-ng-1.5.1/jsonpath_ng/parser.out' mode='w' encoding='UTF-8'> new_parser = ply.yacc.yacc(module=self, ResourceWarning: Enable tracemalloc to get the object allocation traceback ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (sorted_list_indexed) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (sorted_dict) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (sorted_dict_indexed) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (len_list) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (len_dict) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (len_str) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (filter_exists_syntax1) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (filter_exists_syntax2) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (filter_exists_syntax3) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (filter_exists_syntax4) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (filter_eq1) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (filter_eq2) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (filter_eq3) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (filter_gt) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (filter_and) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (filter_float_gt) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (sort1) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (sort1_indexed) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (sort2) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (sort2_indexed) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (sort3) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (sort3_indexed) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (sort4) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (sort4_indexed) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (sort5_twofields) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (sort5_indexed) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (arithmetic_number_only) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (arithmetic_mul1) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (arithmetic_mul2) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (arithmetic_mul3) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (arithmetic_mul4) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (arithmetic_mul5) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (arithmetic_mul6) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (arithmetic_mul7) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (arithmetic_str0) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (arithmetic_str1) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (arithmetic_str2) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (arithmetic_str3) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (arithmetic_str4) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (arithmetic_list1) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (arithmetic_list2) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (arithmetic_list_err1) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (arithmetic_err1) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (arithmetic_err2) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (real_life_example1) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (real_life_example2) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (real_life_example3) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (real_life_example4) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (sub1) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (sub2) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (str1) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (split1) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (split2) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (bug-#2-correct) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (bug-#2-wrong) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (boolean-filter-true) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (boolean-filter-false) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (boolean-filter-other-datatypes-involved) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_fields_value (tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext) tests.test_jsonpath_rw_ext.Testjsonpath_ng_ext.test_fields_value (boolean-filter-string-true-string-literal) ... WARNING [jsonpath_ng.parser] 50 shift/reduce conflicts ok test_atomic (tests.test_parser.TestParser) ... ok test_nested (tests.test_parser.TestParser) ... ok ---------------------------------------------------------------------- Ran 123 tests in 3.453s OK parse("*").find({'foo': 1, 'baz': 2}).paths =?= {'foo', 'baz'} parse("*").find({'foo': 1, 'baz': 2}).paths =?= {'id', 'foo', 'baz'} parse("foo").find({'foo': 'baz'}) =?= ['baz'] parse("foo,baz").find({'foo': 1, 'baz': 2}) =?= [1, 2] parse("@foo").find({'@foo': 1}) =?= [1] parse("*").find({'foo': 1, 'baz': 2}) =?= {1, 2} parse("*").find({'foo': 1, 'baz': 2}) =?= {1, 2, '`this`'} parse("foo."bar-baz"").find({'foo': {'bar-baz': 3}}) =?= [3] parse("foo.["bar-baz","blah-blah"]").find({'foo': {'bar-baz': 3, 'blah-blah': 5}}) =?= [3, 5] parse("[0].id").find([42]) =?= ['[0]'] parse("[2].id").find([34, 65, 29, 59]) =?= ['[2]'] parse("[0]").find([42]).paths =?= ['[0]'] parse("[2]").find([34, 65, 29, 59]).paths =?= ['[2]'] parse("[0]").find([42]) =?= [42] parse("[5]").find([42]) =?= [] parse("[2]").find([34, 65, 29, 59]) =?= [29] parse("foo.baz.`parent`").find({'foo': {'baz': 3}}) =?= [{'baz': 3}] parse("foo.`parent`.foo.baz.`parent`.baz.bizzle").find({'foo': {'baz': {'bizzle': 5}}}) =?= [5] parse("$.id").find({'foo': 'baz'}) =?= ['$'] parse("foo.$.id").find({'foo': 'baz', 'id': 'bizzle'}) =?= ['bizzle'] parse("foo.$.baz.id").find({'foo': 4, 'baz': 3}) =?= ['baz'] parse("$").find({'foo': 'baz'}).paths =?= ['$'] parse("foo.$").find({'foo': 'baz'}).paths =?= ['$'] parse("foo.$.foo").find({'foo': 'baz'}).paths =?= ['foo'] parse("$").find({'foo': 'baz'}) =?= [{'foo': 'baz'}] parse("foo.$").find({'foo': 'baz'}) =?= [{'foo': 'baz'}] parse("foo.$.foo").find({'foo': 'baz'}) =?= ['baz'] parse("[*].id").find([1, 2, 3]) =?= ['[0]', '[1]', '[2]'] parse("[1:].id").find([1, 2, 3, 4]) =?= ['[1]', '[2]', '[3]'] parse("[*]").find([1, 2, 3]).paths =?= ['[0]', '[1]', '[2]'] parse("[1:]").find([1, 2, 3, 4]).paths =?= ['[1]', '[2]', '[3]'] parse("[*]").find([1, 2, 3]) =?= [1, 2, 3] parse("[*]").find(range(1, 4)) =?= [1, 2, 3] parse("[1:]").find([1, 2, 3, 4]) =?= [2, 3, 4] parse("[:2]").find([1, 2, 3, 4]) =?= [1, 2] parse("[*]").find(1) =?= [1] parse("[0:]").find(1) =?= [1] parse("[*]").find({'foo': 1}) =?= [{'foo': 1}] parse("[*].foo").find({'foo': 1}) =?= [1] parse("id").find({'foo': 'baz'}) =?= ['`this`'] parse("foo.`this`.id").find({'foo': 'baz'}) =?= ['foo'] parse("foo.`this`.baz.id").find({'foo': {'baz': 3}}) =?= ['foo.baz'] parse("`this`").find({'foo': 'baz'}).paths =?= ['`this`'] parse("foo.`this`").find({'foo': 'baz'}).paths =?= ['foo'] parse("foo.`this`.baz").find({'foo': {'baz': 3}}).paths =?= ['foo.baz'] parse("`this`").find({'foo': 'baz'}) =?= [{'foo': 'baz'}] parse("foo.`this`").find({'foo': 'baz'}) =?= ['baz'] parse("foo.`this`.baz").find({'foo': {'baz': 3}}) =?= [3] foo =?= foo * =?= * baz,bizzle =?= baz,bizzle [1] =?= [1] [1:] =?= [1] [:] =?= [*] [*] =?= [*] [:2] =?= [:2] [1:2] =?= [1:2] [5:-2] =?= [5:-2] foo.baz =?= foo.baz foo.baz,bizzle =?= foo.baz,bizzle foo where baz =?= foo where baz foo..baz =?= foo..baz foo..baz.bing =?= foo..baz.bing + RPM_EC=0 ++ jobs -p + exit 0 Processing files: python3-jsonpath-ng-1.5.1-6.fc36.noarch Executing(%doc): /bin/sh -e /var/tmp/rpm-tmp.ELTT1f + umask 022 + cd /builddir/build/BUILD + cd jsonpath-ng-1.5.1 + DOCDIR=/builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-6.fc36.x86_64/usr/share/doc/python3-jsonpath-ng + export LC_ALL=C + LC_ALL=C + export DOCDIR + /usr/bin/mkdir -p /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-6.fc36.x86_64/usr/share/doc/python3-jsonpath-ng + cp -pr README.rst /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-6.fc36.x86_64/usr/share/doc/python3-jsonpath-ng + RPM_EC=0 ++ jobs -p + exit 0 Provides: python-jsonpath-ng = 1.5.1-6.fc36 python3-jsonpath-ng = 1.5.1-6.fc36 python3.10-jsonpath-ng = 1.5.1-6.fc36 python3.10dist(jsonpath-ng) = 1.5.1 python3dist(jsonpath-ng) = 1.5.1 Requires(rpmlib): rpmlib(CompressedFileNames) <= 3.0.4-1 rpmlib(FileDigests) <= 4.6.0-1 rpmlib(PartialHardlinkSets) <= 4.0.4-1 rpmlib(PayloadFilesHavePrefix) <= 4.0-1 Requires: /usr/bin/python3 python(abi) = 3.10 python3.10dist(decorator) python3.10dist(ply) python3.10dist(six) Obsoletes: python-jsonpath-ng < 1.5.1-6.fc36 python3.10-jsonpath-ng < 1.5.1-6.fc36 Checking for unpackaged file(s): /usr/lib/rpm/check-files /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-6.fc36.x86_64 Wrote: /builddir/build/RPMS/python3-jsonpath-ng-1.5.1-6.fc36.noarch.rpm Executing(%clean): /bin/sh -e /var/tmp/rpm-tmp.GmVZVt + umask 022 + cd /builddir/build/BUILD + cd jsonpath-ng-1.5.1 + /usr/bin/rm -rf /builddir/build/BUILDROOT/python-jsonpath-ng-1.5.1-6.fc36.x86_64 + RPM_EC=0 ++ jobs -p + exit 0 Child return code was: 0